2025-07-17T08:33:59.2638593Z Current runner version: '2.326.0' 2025-07-17T08:33:59.2643632Z Runner name: 'linux.rocm.gpu.mi300.2-8zrv9-runner-r2whm' 2025-07-17T08:33:59.2644336Z Runner group name: 'default' 2025-07-17T08:33:59.2645009Z Machine name: 'linux' 2025-07-17T08:33:59.2647177Z ##[group]GITHUB_TOKEN Permissions 2025-07-17T08:33:59.2648873Z Contents: read 2025-07-17T08:33:59.2649298Z Metadata: read 2025-07-17T08:33:59.2649838Z ##[endgroup] 2025-07-17T08:33:59.2651664Z Secret source: Actions 2025-07-17T08:33:59.2652299Z Prepare workflow directory 2025-07-17T08:33:59.3041879Z Prepare all required actions 2025-07-17T08:33:59.3071200Z Getting action download info 2025-07-17T08:33:59.6612816Z Download action repository 'pytorch/pytorch@main' (SHA:39ac189808c61588f3594dbc2fc1d69bb6194c47) 2025-07-17T08:34:05.8537955Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-07-17T08:34:06.3096058Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-07-17T08:34:06.7116763Z Download action repository 'pytorch/test-infra@main' (SHA:a9ec424ad5e5851e47d68139cfd953b4031778d5) 2025-07-17T08:34:07.7899186Z Download action repository 'actions/upload-artifact@ea165f8d65b6e75b540449e92b4886f43607fa02' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-07-17T08:34:08.3803893Z Getting action download info 2025-07-17T08:34:08.5267023Z Download action repository 'actions/checkout@v4' (SHA:11bd71901bbe5b1630ceea73d27597364c9af683) 2025-07-17T08:34:08.9599197Z Getting action download info 2025-07-17T08:34:09.0873864Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-07-17T08:34:09.5539115Z Getting action download info 2025-07-17T08:34:09.6880702Z Uses: pytorch/pytorch/.github/workflows/_rocm-test.yml@refs/heads/main (a38f433be2e94a64b095a44ba39879d02d0c2316) 2025-07-17T08:34:09.6883563Z ##[group] Inputs 2025-07-17T08:34:09.6883825Z build-environment: linux-noble-rocm-py3.12-mi300 2025-07-17T08:34:09.6885165Z test-matrix: {"include": [{"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}]} 2025-07-17T08:34:09.6886711Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:34:09.6887166Z sync-tag: 2025-07-17T08:34:09.6887807Z timeout-minutes: 300 2025-07-17T08:34:09.6887991Z tests-to-include: 2025-07-17T08:34:09.6888146Z dashboard-tag: 2025-07-17T08:34:09.6888647Z disable-monitor: true 2025-07-17T08:34:09.6888820Z monitor-log-interval: 5 2025-07-17T08:34:09.6889015Z monitor-data-collect-interval: 1 2025-07-17T08:34:09.6889212Z ##[endgroup] 2025-07-17T08:34:09.6889510Z Complete job name: linux-noble-rocm-py3.12-mi300 / test (default, 2, 6, linux.rocm.gpu.mi300.2, unstable) 2025-07-17T08:34:09.7324172Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-07-17T08:34:09.7324673Z with: 2025-07-17T08:34:09.7324817Z no-sudo: true 2025-07-17T08:34:09.7324975Z submodules: recursive 2025-07-17T08:34:09.7325130Z fetch-depth: 0 2025-07-17T08:34:09.7325381Z env: 2025-07-17T08:34:09.7325527Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:34:09.7325694Z ##[endgroup] 2025-07-17T08:34:09.7392378Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-07-17T08:34:09.7393293Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-07-17T08:34:09.7409024Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:34:09.7409288Z env: 2025-07-17T08:34:09.7409425Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:34:09.7409589Z ##[endgroup] 2025-07-17T08:34:09.7595548Z ##[group]Run # Use all available CPUs for fetching 2025-07-17T08:34:09.7595883Z # Use all available CPUs for fetching 2025-07-17T08:34:09.7596094Z cd "${GITHUB_WORKSPACE}" 2025-07-17T08:34:09.7596307Z git config --global fetch.parallel 0 2025-07-17T08:34:09.7596540Z git config --global submodule.fetchJobs 0 2025-07-17T08:34:09.7596831Z  2025-07-17T08:34:09.7597061Z # Clean workspace. The default checkout action should also do this, but 2025-07-17T08:34:09.7597355Z # do it here as well just in case 2025-07-17T08:34:09.7597592Z if [[ -d .git ]]; then 2025-07-17T08:34:09.7597775Z  if [ -z "${NO_SUDO}" ]; then 2025-07-17T08:34:09.7597968Z  sudo git clean -ffdx 2025-07-17T08:34:09.7598141Z  else 2025-07-17T08:34:09.7598298Z  git clean -ffdx 2025-07-17T08:34:09.7598460Z  fi 2025-07-17T08:34:09.7598605Z fi 2025-07-17T08:34:09.7610650Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:34:09.7610893Z env: 2025-07-17T08:34:09.7611050Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:34:09.7611217Z NO_SUDO: true 2025-07-17T08:34:09.7611370Z ##[endgroup] 2025-07-17T08:34:09.7769176Z ##[group]Run actions/checkout@v4 2025-07-17T08:34:09.7769424Z with: 2025-07-17T08:34:09.7769592Z ref: a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:34:09.7769806Z fetch-depth: 0 2025-07-17T08:34:09.7769959Z submodules: recursive 2025-07-17T08:34:09.7770129Z show-progress: false 2025-07-17T08:34:09.7770311Z repository: pytorch/pytorch 2025-07-17T08:34:09.7770589Z token: *** 2025-07-17T08:34:09.7770738Z ssh-strict: true 2025-07-17T08:34:09.7770885Z ssh-user: git 2025-07-17T08:34:09.7771048Z persist-credentials: true 2025-07-17T08:34:09.7771212Z clean: true 2025-07-17T08:34:09.7771388Z sparse-checkout-cone-mode: true 2025-07-17T08:34:09.7771571Z fetch-tags: false 2025-07-17T08:34:09.7771726Z lfs: false 2025-07-17T08:34:09.7771875Z set-safe-directory: true 2025-07-17T08:34:09.7772062Z env: 2025-07-17T08:34:09.7772202Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:34:09.7772365Z ##[endgroup] 2025-07-17T08:34:09.8911260Z Syncing repository: pytorch/pytorch 2025-07-17T08:34:09.8912238Z ##[group]Getting Git version info 2025-07-17T08:34:09.8912519Z Working directory is '/home/runner/_work/pytorch/pytorch' 2025-07-17T08:34:09.8912902Z [command]/usr/bin/git version 2025-07-17T08:34:09.8921678Z git version 2.50.1 2025-07-17T08:34:09.8942214Z ##[endgroup] 2025-07-17T08:34:09.8950811Z Copying '/home/runner/.gitconfig' to '/home/runner/_work/_temp/c9776a9f-9943-44a8-a517-c327d1b6b520/.gitconfig' 2025-07-17T08:34:09.8964862Z Temporarily overriding HOME='/home/runner/_work/_temp/c9776a9f-9943-44a8-a517-c327d1b6b520' before making global git config changes 2025-07-17T08:34:09.8965377Z Adding repository directory to the temporary git global config as a safe directory 2025-07-17T08:34:09.8969132Z [command]/usr/bin/git config --global --add safe.directory /home/runner/_work/pytorch/pytorch 2025-07-17T08:34:09.8999463Z Deleting the contents of '/home/runner/_work/pytorch/pytorch' 2025-07-17T08:34:09.9002674Z ##[group]Initializing the repository 2025-07-17T08:34:09.9006055Z [command]/usr/bin/git init /home/runner/_work/pytorch/pytorch 2025-07-17T08:34:09.9062472Z hint: Using 'master' as the name for the initial branch. This default branch name 2025-07-17T08:34:09.9062843Z hint: is subject to change. To configure the initial branch name to use in all 2025-07-17T08:34:09.9063178Z hint: of your new repositories, which will suppress this warning, call: 2025-07-17T08:34:09.9063703Z hint: 2025-07-17T08:34:09.9063906Z hint: git config --global init.defaultBranch 2025-07-17T08:34:09.9064125Z hint: 2025-07-17T08:34:09.9064336Z hint: Names commonly chosen instead of 'master' are 'main', 'trunk' and 2025-07-17T08:34:09.9064664Z hint: 'development'. The just-created branch can be renamed via this command: 2025-07-17T08:34:09.9064922Z hint: 2025-07-17T08:34:09.9065065Z hint: git branch -m 2025-07-17T08:34:09.9065228Z hint: 2025-07-17T08:34:09.9065447Z hint: Disable this message with "git config set advice.defaultBranchName false" 2025-07-17T08:34:09.9065809Z Initialized empty Git repository in /home/runner/_work/pytorch/pytorch/.git/ 2025-07-17T08:34:09.9073121Z [command]/usr/bin/git remote add origin https://github.com/pytorch/pytorch 2025-07-17T08:34:09.9113196Z ##[endgroup] 2025-07-17T08:34:09.9113502Z ##[group]Disabling automatic garbage collection 2025-07-17T08:34:09.9116646Z [command]/usr/bin/git config --local gc.auto 0 2025-07-17T08:34:09.9143416Z ##[endgroup] 2025-07-17T08:34:09.9145648Z ##[group]Setting up auth 2025-07-17T08:34:09.9149107Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-07-17T08:34:09.9182697Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-07-17T08:34:09.9479799Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-07-17T08:34:09.9506106Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-07-17T08:34:09.9798806Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-07-17T08:34:09.9839422Z ##[endgroup] 2025-07-17T08:34:09.9839763Z ##[group]Fetching the repository 2025-07-17T08:34:09.9853312Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-07-17T08:34:49.8720539Z From https://github.com/pytorch/pytorch 2025-07-17T08:34:49.8721127Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-07-17T08:34:49.8723499Z * [new branch] 20250616_dtype_docs -> origin/20250616_dtype_docs 2025-07-17T08:34:49.8726519Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-07-17T08:34:49.8727683Z * [new branch] JackCaoG/dynamo_make_fx_non_core_aten_ops -> origin/JackCaoG/dynamo_make_fx_non_core_aten_ops 2025-07-17T08:34:49.8731105Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-07-17T08:34:49.8734523Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-07-17T08:34:49.8737886Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-07-17T08:34:49.8741101Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-07-17T08:34:49.8744432Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-07-17T08:34:49.8747727Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-07-17T08:34:49.8751781Z * [new branch] ZainRizvi-patch-1 -> origin/ZainRizvi-patch-1 2025-07-17T08:34:49.8754822Z * [new branch] add-missing-args-normalization -> origin/add-missing-args-normalization 2025-07-17T08:34:49.8758138Z * [new branch] addUtilForLinuxBuild -> origin/addUtilForLinuxBuild 2025-07-17T08:34:49.8761356Z * [new branch] add_windows_testing_back -> origin/add_windows_testing_back 2025-07-17T08:34:49.8764824Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-07-17T08:34:49.8768199Z * [new branch] addsimde -> origin/addsimde 2025-07-17T08:34:49.8772660Z * [new branch] adi/test -> origin/adi/test 2025-07-17T08:34:49.8776041Z * [new branch] adi/test_fusions -> origin/adi/test_fusions 2025-07-17T08:34:49.8779487Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-07-17T08:34:49.8782785Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-07-17T08:34:49.8786164Z * [new branch] adi/update_openblas -> origin/adi/update_openblas 2025-07-17T08:34:49.8791615Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-07-17T08:34:49.8795235Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-07-17T08:34:49.8798694Z * [new branch] albanD-patch-1 -> origin/albanD-patch-1 2025-07-17T08:34:49.8802003Z * [new branch] alt-disable -> origin/alt-disable 2025-07-17T08:34:49.8806752Z * [new branch] angelayi/155426 -> origin/angelayi/155426 2025-07-17T08:34:49.8809919Z * [new branch] angelayi/157183 -> origin/angelayi/157183 2025-07-17T08:34:49.8813344Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-07-17T08:34:49.8816838Z * [new branch] angelayi/aoti_custom_op -> origin/angelayi/aoti_custom_op 2025-07-17T08:34:49.8820058Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-07-17T08:34:49.8823345Z * [new branch] angelayi/benchmark2 -> origin/angelayi/benchmark2 2025-07-17T08:34:49.8826568Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-07-17T08:34:49.8830007Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-07-17T08:34:49.8833274Z * [new branch] angelayi/customop -> origin/angelayi/customop 2025-07-17T08:34:49.8836620Z * [new branch] angelayi/del_lib -> origin/angelayi/del_lib 2025-07-17T08:34:49.8839843Z * [new branch] angelayi/docs -> origin/angelayi/docs 2025-07-17T08:34:49.8843191Z * [new branch] angelayi/docs2 -> origin/angelayi/docs2 2025-07-17T08:34:49.8846594Z * [new branch] angelayi/dynamo_fake_input -> origin/angelayi/dynamo_fake_input 2025-07-17T08:34:49.8872077Z * [new branch] angelayi/errmsg -> origin/angelayi/errmsg 2025-07-17T08:34:49.8872542Z * [new branch] angelayi/fake_device -> origin/angelayi/fake_device 2025-07-17T08:34:49.8872918Z * [new branch] angelayi/int_lift_constants -> origin/angelayi/int_lift_constants 2025-07-17T08:34:49.8873291Z * [new branch] angelayi/logging.bak -> origin/angelayi/logging.bak 2025-07-17T08:34:49.8873625Z * [new branch] angelayi/logging2 -> origin/angelayi/logging2 2025-07-17T08:34:49.8874030Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-07-17T08:34:49.8874352Z * [new branch] angelayi/pytree -> origin/angelayi/pytree 2025-07-17T08:34:49.8874942Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-07-17T08:34:49.8877573Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-07-17T08:34:49.8880825Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-07-17T08:34:49.8883172Z * [new branch] angelayi/torchgenooops -> origin/angelayi/torchgenooops 2025-07-17T08:34:49.8886789Z * [new branch] angelayi/update_schema_msg -> origin/angelayi/update_schema_msg 2025-07-17T08:34:49.8890063Z * [new branch] aoti_static_linkage -> origin/aoti_static_linkage 2025-07-17T08:34:49.8893474Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-07-17T08:34:49.8896685Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-07-17T08:34:49.8899838Z * [new branch] atalman-patch-1 -> origin/atalman-patch-1 2025-07-17T08:34:49.8903236Z * [new branch] atalman-patch-2 -> origin/atalman-patch-2 2025-07-17T08:34:49.8906622Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-07-17T08:34:49.8909867Z * [new branch] atalman-patch-4 -> origin/atalman-patch-4 2025-07-17T08:34:49.8913175Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-07-17T08:34:49.8916594Z * [new branch] atalman-patch-6 -> origin/atalman-patch-6 2025-07-17T08:34:49.8919843Z * [new branch] atalman-patch-7 -> origin/atalman-patch-7 2025-07-17T08:34:49.8923108Z * [new branch] atalman-patch-8 -> origin/atalman-patch-8 2025-07-17T08:34:49.8926487Z * [new branch] atalman_inductor_2.3.0 -> origin/atalman_inductor_2.3.0 2025-07-17T08:34:49.8929679Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-07-17T08:34:49.8932960Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-07-17T08:34:49.8936234Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-07-17T08:34:49.8940737Z * [new branch] base/1.5 -> origin/base/1.5 2025-07-17T08:34:49.8944221Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-07-17T08:34:49.8947576Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-07-17T08:34:49.8950846Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-07-17T08:34:49.8955698Z * [new branch] benjaminglass1/mark-large-tensor-tests-serial -> origin/benjaminglass1/mark-large-tensor-tests-serial 2025-07-17T08:34:49.8960089Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-07-17T08:34:49.8964498Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-07-17T08:34:49.8968884Z * [new branch] bf/cg-capture-size -> origin/bf/cg-capture-size 2025-07-17T08:34:49.8972140Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-07-17T08:34:49.8975666Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-07-17T08:34:49.8978940Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-07-17T08:34:49.8982990Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-07-17T08:34:49.8986290Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-07-17T08:34:49.8989624Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-07-17T08:34:49.8993225Z * [new branch] bf/kernel-benchmark -> origin/bf/kernel-benchmark 2025-07-17T08:34:49.8996990Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-07-17T08:34:49.9000499Z * [new branch] bf/partition-turn-on -> origin/bf/partition-turn-on 2025-07-17T08:34:49.9003911Z * [new branch] bf/recompile-context -> origin/bf/recompile-context 2025-07-17T08:34:49.9007358Z * [new branch] bf/reduce-scatter-copy-in -> origin/bf/reduce-scatter-copy-in 2025-07-17T08:34:49.9010749Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-07-17T08:34:49.9013870Z * [new branch] bf16adamw -> origin/bf16adamw 2025-07-17T08:34:49.9017291Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-07-17T08:34:49.9020525Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-07-17T08:34:49.9023856Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-07-17T08:34:49.9026956Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-07-17T08:34:49.9030133Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-07-17T08:34:49.9033324Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-07-17T08:34:49.9036778Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-07-17T08:34:49.9039970Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-07-17T08:34:49.9043121Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-07-17T08:34:49.9046387Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-07-17T08:34:49.9049634Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-07-17T08:34:49.9052879Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-07-17T08:34:49.9056176Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-07-17T08:34:49.9059646Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-07-17T08:34:49.9062819Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-07-17T08:34:49.9066007Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-07-17T08:34:49.9069278Z * [new branch] bit8_transpose_pack -> origin/bit8_transpose_pack 2025-07-17T08:34:49.9073832Z * [new branch] bowbao/bench_updates_stage -> origin/bowbao/bench_updates_stage 2025-07-17T08:34:49.9077309Z * [new branch] bowbao/dort_rewriter -> origin/bowbao/dort_rewriter 2025-07-17T08:34:49.9080538Z * [new branch] bowbao/wip_prs -> origin/bowbao/wip_prs 2025-07-17T08:34:49.9084962Z * [new branch] bowenbao/partial_min_max_reduce -> origin/bowenbao/partial_min_max_reduce 2025-07-17T08:34:49.9089395Z * [new branch] brister/always_wrapper_ir -> origin/brister/always_wrapper_ir 2025-07-17T08:34:49.9092672Z * [new branch] brister/test_block_ptr_same -> origin/brister/test_block_ptr_same 2025-07-17T08:34:49.9096012Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-07-17T08:34:49.9099525Z * [new branch] brister/wrapper_fx_precomputed_size -> origin/brister/wrapper_fx_precomputed_size 2025-07-17T08:34:49.9102550Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-07-17T08:34:49.9105946Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-07-17T08:34:49.9111395Z * [new branch] camyll/revert-94bc900da97ad7f3c35b3b819bb53b23c74b581a-for-release-2.8 -> origin/camyll/revert-94bc900da97ad7f3c35b3b819bb53b23c74b581a-for-release-2.8 2025-07-17T08:34:49.9114752Z * [new branch] camyll/test_precommit_hooks_lintrunner -> origin/camyll/test_precommit_hooks_lintrunner 2025-07-17T08:34:49.9119308Z * [new branch] camyllh/cherry-pick-155109-revert-to-release2.8 -> origin/camyllh/cherry-pick-155109-revert-to-release2.8 2025-07-17T08:34:49.9122583Z * [new branch] camyllh/cherrypick-151547-for-release28 -> origin/camyllh/cherrypick-151547-for-release28 2025-07-17T08:34:49.9125966Z * [new branch] camyllh/cherrypick_152932_for_release_2.8 -> origin/camyllh/cherrypick_152932_for_release_2.8 2025-07-17T08:34:49.9129157Z * [new branch] camyllh/revert-138222-in-release-2.8 -> origin/camyllh/revert-138222-in-release-2.8 2025-07-17T08:34:49.9132418Z * [new branch] camyllh/revert-155412-on-release-branch -> origin/camyllh/revert-155412-on-release-branch 2025-07-17T08:34:49.9135577Z * [new branch] camyllh/revert-156552-on-release-2.8 -> origin/camyllh/revert-156552-on-release-2.8 2025-07-17T08:34:49.9138850Z * [new branch] camyllh/revert_156517_on_release_2.8 -> origin/camyllh/revert_156517_on_release_2.8 2025-07-17T08:34:49.9142233Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-07-17T08:34:49.9145705Z * [new branch] cherry-pick-147095-by-pytorch_bot_bot_ -> origin/cherry-pick-147095-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9149238Z * [new branch] cherry-pick-149654-by-pytorch_bot_bot_ -> origin/cherry-pick-149654-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9152491Z * [new branch] cherry-pick-151939-by-pytorch_bot_bot_ -> origin/cherry-pick-151939-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9156036Z * [new branch] cherry-pick-154174-by-pytorch_bot_bot_ -> origin/cherry-pick-154174-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9159229Z * [new branch] cherry-pick-155214-by-pytorch_bot_bot_ -> origin/cherry-pick-155214-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9162951Z * [new branch] cherry-pick-155255-by-pytorch_bot_bot_ -> origin/cherry-pick-155255-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9166187Z * [new branch] cherry-pick-155896-by-pytorch_bot_bot_ -> origin/cherry-pick-155896-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9169434Z * [new branch] cherry-pick-156260-by-pytorch_bot_bot_ -> origin/cherry-pick-156260-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9172920Z * [new branch] cherry-pick-156499-by-pytorch_bot_bot_ -> origin/cherry-pick-156499-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9176202Z * [new branch] cherry-pick-156664-by-pytorch_bot_bot_ -> origin/cherry-pick-156664-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9179444Z * [new branch] cherry-pick-156708-by-pytorch_bot_bot_ -> origin/cherry-pick-156708-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9182789Z * [new branch] cherry-pick-156719-by-pytorch_bot_bot_ -> origin/cherry-pick-156719-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9186043Z * [new branch] cherry-pick-156731-by-pytorch_bot_bot_ -> origin/cherry-pick-156731-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9189208Z * [new branch] cherry-pick-156783-by-pytorch_bot_bot_ -> origin/cherry-pick-156783-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9192481Z * [new branch] cherry-pick-156847-by-pytorch_bot_bot_ -> origin/cherry-pick-156847-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9196124Z * [new branch] cherry-pick-156888-by-pytorch_bot_bot_ -> origin/cherry-pick-156888-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9199254Z * [new branch] cherry-pick-156898-by-pytorch_bot_bot_ -> origin/cherry-pick-156898-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9202947Z * [new branch] cherry-pick-156910-by-pytorch_bot_bot_ -> origin/cherry-pick-156910-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9206211Z * [new branch] cherry-pick-157014-by-pytorch_bot_bot_ -> origin/cherry-pick-157014-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9209510Z * [new branch] cherry-pick-157017-by-pytorch_bot_bot_ -> origin/cherry-pick-157017-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9212862Z * [new branch] cherry-pick-157023-by-pytorch_bot_bot_ -> origin/cherry-pick-157023-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9216075Z * [new branch] cherry-pick-157130-by-pytorch_bot_bot_ -> origin/cherry-pick-157130-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9219383Z * [new branch] cherry-pick-157292-by-pytorch_bot_bot_ -> origin/cherry-pick-157292-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9222598Z * [new branch] cherry-pick-157322-by-pytorch_bot_bot_ -> origin/cherry-pick-157322-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9226049Z * [new branch] cherry-pick-157453-by-pytorch_bot_bot_ -> origin/cherry-pick-157453-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9229309Z * [new branch] cherry-pick-157466-by-pytorch_bot_bot_ -> origin/cherry-pick-157466-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9232729Z * [new branch] cherry-pick-157467-by-pytorch_bot_bot_ -> origin/cherry-pick-157467-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9236216Z * [new branch] cherry-pick-157503-by-pytorch_bot_bot_ -> origin/cherry-pick-157503-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9239369Z * [new branch] cherry-pick-157513-by-pytorch_bot_bot_ -> origin/cherry-pick-157513-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9242561Z * [new branch] cherry-pick-157516-by-pytorch_bot_bot_ -> origin/cherry-pick-157516-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9245820Z * [new branch] cherry-pick-157558-by-pytorch_bot_bot_ -> origin/cherry-pick-157558-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9249060Z * [new branch] cherry-pick-157598-by-pytorch_bot_bot_ -> origin/cherry-pick-157598-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9252400Z * [new branch] cherry-pick-157600-by-pytorch_bot_bot_ -> origin/cherry-pick-157600-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9255820Z * [new branch] cherry-pick-157630-by-pytorch_bot_bot_ -> origin/cherry-pick-157630-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9259124Z * [new branch] cherry-pick-157695-by-pytorch_bot_bot_ -> origin/cherry-pick-157695-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9262382Z * [new branch] cherry-pick-157732-by-pytorch_bot_bot_ -> origin/cherry-pick-157732-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9265657Z * [new branch] cherry-pick-157733-by-pytorch_bot_bot_ -> origin/cherry-pick-157733-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9268959Z * [new branch] cherry-pick-157843-by-pytorch_bot_bot_ -> origin/cherry-pick-157843-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9272370Z * [new branch] cherry-pick-158064-by-pytorch_bot_bot_ -> origin/cherry-pick-158064-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9275841Z * [new branch] cherry-pick-158126-by-pytorch_bot_bot_ -> origin/cherry-pick-158126-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9279151Z * [new branch] cherry-pick-158152-by-pytorch_bot_bot_ -> origin/cherry-pick-158152-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9282611Z * [new branch] cherry-pick-158301-by-pytorch_bot_bot_ -> origin/cherry-pick-158301-by-pytorch_bot_bot_ 2025-07-17T08:34:49.9286891Z * [new branch] chilli/flex_vllm -> origin/chilli/flex_vllm 2025-07-17T08:34:49.9290167Z * [new branch] ck-epilogue-fix -> origin/ck-epilogue-fix 2025-07-17T08:34:49.9293464Z * [new branch] ckluk2-compileThread-1 -> origin/ckluk2-compileThread-1 2025-07-17T08:34:49.9296741Z * [new branch] ckluk2-compileThread-2 -> origin/ckluk2-compileThread-2 2025-07-17T08:34:49.9300041Z * [new branch] ckluk2-compileThread-64 -> origin/ckluk2-compileThread-64 2025-07-17T08:34:49.9303036Z * [new branch] ckluk2-test-1 -> origin/ckluk2-test-1 2025-07-17T08:34:49.9306341Z * [new branch] cleantest1 -> origin/cleantest1 2025-07-17T08:34:49.9311538Z * [new branch] codex/create-test-for-tensor-memory-leak-in-cudagraph -> origin/codex/create-test-for-tensor-memory-leak-in-cudagraph 2025-07-17T08:34:49.9315219Z * [new branch] codex/remove-trigger_azure_pipeline.py-and-references -> origin/codex/remove-trigger_azure_pipeline.py-and-references 2025-07-17T08:34:49.9318693Z * [new branch] codex/update-default-size_hint-fallback-value -> origin/codex/update-default-size_hint-fallback-value 2025-07-17T08:34:49.9322167Z * [new branch] codex/update-size_hint-api-to-handle-unbacked-symints -> origin/codex/update-size_hint-api-to-handle-unbacked-symints 2025-07-17T08:34:49.9325152Z * [new branch] collective_c_shim -> origin/collective_c_shim 2025-07-17T08:34:49.9328406Z * [new branch] compile_fsdp2_disable_stream_and_event -> origin/compile_fsdp2_disable_stream_and_event 2025-07-17T08:34:49.9331569Z * [new branch] comply-with-setuptools -> origin/comply-with-setuptools 2025-07-17T08:34:49.9334794Z * [new branch] context_test -> origin/context_test 2025-07-17T08:34:49.9338125Z * [new branch] convert-complex-numbers -> origin/convert-complex-numbers 2025-07-17T08:34:49.9342469Z * [new branch] copilot/fix-157446 -> origin/copilot/fix-157446 2025-07-17T08:34:49.9345746Z * [new branch] copy_graph -> origin/copy_graph 2025-07-17T08:34:49.9350174Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-07-17T08:34:49.9354831Z * [new branch] csl/add_file_merge_conflict_csv -> origin/csl/add_file_merge_conflict_csv 2025-07-17T08:34:49.9358131Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-07-17T08:34:49.9361421Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-07-17T08:34:49.9364634Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-07-17T08:34:49.9367978Z * [new branch] csl/ci_upload_artifacts_with_running -> origin/csl/ci_upload_artifacts_with_running 2025-07-17T08:34:49.9371239Z * [new branch] csl/clean_up_max_jobs_override -> origin/csl/clean_up_max_jobs_override 2025-07-17T08:34:49.9374494Z * [new branch] csl/clean_up_more_tags -> origin/csl/clean_up_more_tags 2025-07-17T08:34:49.9377775Z * [new branch] csl/disable_periodic_tests -> origin/csl/disable_periodic_tests 2025-07-17T08:34:49.9380972Z * [new branch] csl/docker_instru_ssh -> origin/csl/docker_instru_ssh 2025-07-17T08:34:49.9384193Z * [new branch] csl/faketensortest_super_setup -> origin/csl/faketensortest_super_setup 2025-07-17T08:34:49.9387409Z * [new branch] csl/fix_docker_imgs_for_release -> origin/csl/fix_docker_imgs_for_release 2025-07-17T08:34:49.9390651Z * [new branch] csl/fix_flaky_dist_test -> origin/csl/fix_flaky_dist_test 2025-07-17T08:34:49.9393853Z * [new branch] csl/fix_rocm_mem -> origin/csl/fix_rocm_mem 2025-07-17T08:34:49.9397314Z * [new branch] csl/katex -> origin/csl/katex 2025-07-17T08:34:49.9400578Z * [new branch] csl/keep_going_main -> origin/csl/keep_going_main 2025-07-17T08:34:49.9404253Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-07-17T08:34:49.9407491Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-07-17T08:34:49.9411041Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-07-17T08:34:49.9414063Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-07-17T08:34:49.9417469Z * [new branch] csl/per_proc_mem -> origin/csl/per_proc_mem 2025-07-17T08:34:49.9420713Z * [new branch] csl/print_set_output -> origin/csl/print_set_output 2025-07-17T08:34:49.9423999Z * [new branch] csl/pytest_timeout -> origin/csl/pytest_timeout 2025-07-17T08:34:49.9427369Z * [new branch] csl/rerun_exit_error -> origin/csl/rerun_exit_error 2025-07-17T08:34:49.9430667Z * [new branch] csl/reuse_old_whl_renames -> origin/csl/reuse_old_whl_renames 2025-07-17T08:34:49.9433954Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-07-17T08:34:49.9437304Z * [new branch] csl/rocm_binary -> origin/csl/rocm_binary 2025-07-17T08:34:49.9440530Z * [new branch] csl/rocm_inductor -> origin/csl/rocm_inductor 2025-07-17T08:34:49.9444081Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-07-17T08:34:49.9447270Z * [new branch] csl/td_dynamo -> origin/csl/td_dynamo 2025-07-17T08:34:49.9450629Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-07-17T08:34:49.9453891Z * [new branch] csl/update_mobile_build_docker_img -> origin/csl/update_mobile_build_docker_img 2025-07-17T08:34:49.9457417Z * [new branch] csl/viable_strict_concurrency_group -> origin/csl/viable_strict_concurrency_group 2025-07-17T08:34:49.9460579Z * [new branch] cublasltrelax2 -> origin/cublasltrelax2 2025-07-17T08:34:49.9463779Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-07-17T08:34:49.9467081Z * [new branch] cudnnsdparefactor -> origin/cudnnsdparefactor 2025-07-17T08:34:49.9470285Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-07-17T08:34:49.9474771Z * [new branch] d4l3k/delete_hook -> origin/d4l3k/delete_hook 2025-07-17T08:34:49.9478000Z * [new branch] d4l3k/dist_queue -> origin/d4l3k/dist_queue 2025-07-17T08:34:49.9481171Z * [new branch] d4l3k/fix_h100_ci -> origin/d4l3k/fix_h100_ci 2025-07-17T08:34:49.9484391Z * [new branch] d4l3k/wait_stream -> origin/d4l3k/wait_stream 2025-07-17T08:34:49.9487715Z * [new branch] dcp_zoc -> origin/dcp_zoc 2025-07-17T08:34:49.9491004Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-07-17T08:34:49.9498557Z * [new branch] dependabot/pip/dot-ci/docker/protobuf-5.29.5 -> origin/dependabot/pip/dot-ci/docker/protobuf-5.29.5 2025-07-17T08:34:49.9505187Z * [new branch] dependabot/pip/tools/build/bazel/requests-2.32.4 -> origin/dependabot/pip/tools/build/bazel/requests-2.32.4 2025-07-17T08:34:49.9508287Z * [new branch] deprecation_decorator -> origin/deprecation_decorator 2025-07-17T08:34:49.9512702Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-07-17T08:34:49.9516274Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-07-17T08:34:49.9521751Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-07-17T08:34:49.9525223Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-07-17T08:34:49.9528772Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-07-17T08:34:49.9532202Z * [new branch] dev/joona/cat_remove_graph -> origin/dev/joona/cat_remove_graph 2025-07-17T08:34:49.9535708Z * [new branch] dev/joona/cumsumerrmsg -> origin/dev/joona/cumsumerrmsg 2025-07-17T08:34:49.9538942Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-07-17T08:34:49.9542552Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-07-17T08:34:49.9546129Z * [new branch] dev/joona/maxpool2dwithindices_errmsg -> origin/dev/joona/maxpool2dwithindices_errmsg 2025-07-17T08:34:49.9549652Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-07-17T08:34:49.9553067Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-07-17T08:34:49.9556880Z * [new branch] dev/joona/sdpa_nans -> origin/dev/joona/sdpa_nans 2025-07-17T08:34:49.9560285Z * [new branch] dev/joona/synchronize_benchmark -> origin/dev/joona/synchronize_benchmark 2025-07-17T08:34:49.9563682Z * [new branch] dev/joona/topk_newapi -> origin/dev/joona/topk_newapi 2025-07-17T08:34:49.9567176Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-07-17T08:34:49.9570618Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-07-17T08:34:49.9573904Z * [new branch] disable -> origin/disable 2025-07-17T08:34:49.9577058Z * [new branch] divup -> origin/divup 2025-07-17T08:34:49.9580515Z * [new branch] divyanshk-log-api-usage-datapipes-1 -> origin/divyanshk-log-api-usage-datapipes-1 2025-07-17T08:34:49.9583892Z * [new branch] e2e-baseline -> origin/e2e-baseline 2025-07-17T08:34:49.9588374Z * [new branch] embg/test_inductor_ci_128B -> origin/embg/test_inductor_ci_128B 2025-07-17T08:34:49.9591822Z * [new branch] embg/test_inductor_ci_base -> origin/embg/test_inductor_ci_base 2025-07-17T08:34:49.9595404Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-07-17T08:34:49.9598668Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-07-17T08:34:49.9601815Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-07-17T08:34:49.9604998Z * [new branch] enable-b200-benchmark -> origin/enable-b200-benchmark 2025-07-17T08:34:49.9608175Z * [new branch] enable_td_rocm -> origin/enable_td_rocm 2025-07-17T08:34:49.9611444Z * [new branch] eqy-patch-1 -> origin/eqy-patch-1 2025-07-17T08:34:49.9614611Z * [new branch] eqy-patch-10 -> origin/eqy-patch-10 2025-07-17T08:34:49.9617917Z * [new branch] example-convert-torch.nn -> origin/example-convert-torch.nn 2025-07-17T08:34:49.9622312Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-07-17T08:34:49.9625626Z * [new branch] exclamaforte/bump-transformer-version -> origin/exclamaforte/bump-transformer-version 2025-07-17T08:34:49.9628875Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-07-17T08:34:49.9632199Z * [new branch] exclamaforte/debug-autotuner-profile -> origin/exclamaforte/debug-autotuner-profile 2025-07-17T08:34:49.9635537Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-07-17T08:34:49.9638895Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-07-17T08:34:49.9642395Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-07-17T08:34:49.9645729Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-07-17T08:34:49.9649000Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-07-17T08:34:49.9652136Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-07-17T08:34:49.9655369Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-07-17T08:34:49.9658616Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-07-17T08:34:49.9662036Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-07-17T08:34:49.9665113Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-07-17T08:34:49.9668513Z * [new branch] exclamaforte/heuristic-choices -> origin/exclamaforte/heuristic-choices 2025-07-17T08:34:49.9671749Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-07-17T08:34:49.9675039Z * [new branch] exclamaforte/log_mul -> origin/exclamaforte/log_mul 2025-07-17T08:34:49.9678317Z * [new branch] exclamaforte/memory-counter -> origin/exclamaforte/memory-counter 2025-07-17T08:34:49.9681481Z * [new branch] exclamaforte/scheduler-refactor -> origin/exclamaforte/scheduler-refactor 2025-07-17T08:34:49.9684875Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-07-17T08:34:49.9688229Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-07-17T08:34:49.9691473Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-07-17T08:34:49.9696292Z * [new branch] exclamforte/gemm-model-final -> origin/exclamforte/gemm-model-final 2025-07-17T08:34:49.9699531Z * [new branch] exec -> origin/exec 2025-07-17T08:34:49.9702894Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-07-17T08:34:49.9706164Z * [new branch] export-D58091437 -> origin/export-D58091437 2025-07-17T08:34:49.9709506Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-07-17T08:34:49.9712821Z * [new branch] export-D68846308 -> origin/export-D68846308 2025-07-17T08:34:49.9718338Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-07-17T08:34:49.9721353Z * [new branch] export-D71446522 -> origin/export-D71446522 2025-07-17T08:34:49.9724602Z * [new branch] export-D72391401 -> origin/export-D72391401 2025-07-17T08:34:49.9727961Z * [new branch] export-D72483950 -> origin/export-D72483950 2025-07-17T08:34:49.9731409Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-07-17T08:34:49.9734579Z * [new branch] export-D73287751 -> origin/export-D73287751 2025-07-17T08:34:49.9737800Z * [new branch] export-D75183591 -> origin/export-D75183591 2025-07-17T08:34:49.9741112Z * [new branch] export-D75605373 -> origin/export-D75605373 2025-07-17T08:34:49.9744298Z * [new branch] export-D75617432 -> origin/export-D75617432 2025-07-17T08:34:49.9747678Z * [new branch] export-D75659965 -> origin/export-D75659965 2025-07-17T08:34:49.9750918Z * [new branch] export-D76080931 -> origin/export-D76080931 2025-07-17T08:34:49.9754549Z * [new branch] export-D76628834 -> origin/export-D76628834 2025-07-17T08:34:49.9757924Z * [new branch] export-D76630499 -> origin/export-D76630499 2025-07-17T08:34:49.9761042Z * [new branch] export-D76797250 -> origin/export-D76797250 2025-07-17T08:34:49.9764499Z * [new branch] export-D76830552 -> origin/export-D76830552 2025-07-17T08:34:49.9767554Z * [new branch] export-D76885271 -> origin/export-D76885271 2025-07-17T08:34:49.9770724Z * [new branch] export-D76885620 -> origin/export-D76885620 2025-07-17T08:34:49.9774041Z * [new branch] export-D76936623 -> origin/export-D76936623 2025-07-17T08:34:49.9777285Z * [new branch] export-D76958268 -> origin/export-D76958268 2025-07-17T08:34:49.9780488Z * [new branch] export-D77639021 -> origin/export-D77639021 2025-07-17T08:34:49.9783760Z * [new branch] export-D78363609 -> origin/export-D78363609 2025-07-17T08:34:49.9786889Z * [new branch] export-D78375400 -> origin/export-D78375400 2025-07-17T08:34:49.9790082Z * [new branch] export-D78431075 -> origin/export-D78431075 2025-07-17T08:34:49.9793265Z * [new branch] export-D78431305 -> origin/export-D78431305 2025-07-17T08:34:49.9796746Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-07-17T08:34:49.9801167Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-07-17T08:34:49.9804582Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-07-17T08:34:49.9807761Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-07-17T08:34:49.9812299Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-07-17T08:34:49.9815619Z * [new branch] fca -> origin/fca 2025-07-17T08:34:49.9818773Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-07-17T08:34:49.9821932Z * [new branch] fca5 -> origin/fca5 2025-07-17T08:34:49.9826490Z * [new branch] fengyuan/external-proj -> origin/fengyuan/external-proj 2025-07-17T08:34:49.9829867Z * [new branch] fengyuan/out-of-tree-xpu-ops-improve-test -> origin/fengyuan/out-of-tree-xpu-ops-improve-test 2025-07-17T08:34:49.9833057Z * [new branch] fengyuan/out-of-tree-xpu-ops-remove-dtype -> origin/fengyuan/out-of-tree-xpu-ops-remove-dtype 2025-07-17T08:34:49.9836432Z * [new branch] fengyuan/test-xpu -> origin/fengyuan/test-xpu 2025-07-17T08:34:49.9840083Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-07-17T08:34:49.9843349Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-07-17T08:34:49.9847890Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-07-17T08:34:49.9851243Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-07-17T08:34:49.9854568Z * [new branch] findhao/fix-indirect-access -> origin/findhao/fix-indirect-access 2025-07-17T08:34:49.9857729Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-07-17T08:34:49.9860950Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-07-17T08:34:49.9864138Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-07-17T08:34:49.9867466Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-07-17T08:34:49.9870697Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-07-17T08:34:49.9874074Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-07-17T08:34:49.9877477Z * [new branch] fix -> origin/fix 2025-07-17T08:34:49.9880641Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-07-17T08:34:49.9883992Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-07-17T08:34:49.9887067Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-07-17T08:34:49.9890306Z * [new branch] fix-inductor-periodic-0528 -> origin/fix-inductor-periodic-0528 2025-07-17T08:34:49.9893624Z * [new branch] fix-issue-144701-attempt-4 -> origin/fix-issue-144701-attempt-4 2025-07-17T08:34:49.9897054Z * [new branch] fix-issue-148701-attempt-1 -> origin/fix-issue-148701-attempt-1 2025-07-17T08:34:49.9900312Z * [new branch] fix-issue-149534-attempt-1 -> origin/fix-issue-149534-attempt-1 2025-07-17T08:34:49.9903558Z * [new branch] fix-rlease-feature-template -> origin/fix-rlease-feature-template 2025-07-17T08:34:49.9906657Z * [new branch] fix_153389 -> origin/fix_153389 2025-07-17T08:34:49.9909953Z * [new branch] fix_allow_train_eval_msg -> origin/fix_allow_train_eval_msg 2025-07-17T08:34:49.9913131Z * [new branch] fix_fake_tensor_issue -> origin/fix_fake_tensor_issue 2025-07-17T08:34:49.9916473Z * [new branch] fix_sbgemm_bf16 -> origin/fix_sbgemm_bf16 2025-07-17T08:34:49.9919811Z * [new branch] fixes-triage -> origin/fixes-triage 2025-07-17T08:34:49.9923213Z * [new branch] flex-lowering -> origin/flex-lowering 2025-07-17T08:34:49.9926475Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-07-17T08:34:49.9930940Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-07-17T08:34:49.9934031Z * [new branch] fsdp-qps-drop -> origin/fsdp-qps-drop 2025-07-17T08:34:49.9937564Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-07-17T08:34:49.9940817Z * [new branch] fsdpv2_3d -> origin/fsdpv2_3d 2025-07-17T08:34:49.9944156Z * [new branch] fsdpv2_3d_m1 -> origin/fsdpv2_3d_m1 2025-07-17T08:34:49.9947649Z * [new branch] fused_qkv -> origin/fused_qkv 2025-07-17T08:34:49.9950954Z * [new branch] fx_cpp -> origin/fx_cpp 2025-07-17T08:34:49.9955758Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-07-17T08:34:49.9962459Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-07-17T08:34:49.9965699Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-07-17T08:34:49.9971157Z * [new branch] gh/CaoE/1/base -> origin/gh/CaoE/1/base 2025-07-17T08:34:49.9974455Z * [new branch] gh/CaoE/1/head -> origin/gh/CaoE/1/head 2025-07-17T08:34:49.9977737Z * [new branch] gh/CaoE/1/orig -> origin/gh/CaoE/1/orig 2025-07-17T08:34:49.9982268Z * [new branch] gh/CaoE/2/base -> origin/gh/CaoE/2/base 2025-07-17T08:34:49.9985677Z * [new branch] gh/CaoE/2/head -> origin/gh/CaoE/2/head 2025-07-17T08:34:49.9989094Z * [new branch] gh/CaoE/2/orig -> origin/gh/CaoE/2/orig 2025-07-17T08:34:49.9993628Z * [new branch] gh/CaoE/3/base -> origin/gh/CaoE/3/base 2025-07-17T08:34:49.9997200Z * [new branch] gh/CaoE/3/head -> origin/gh/CaoE/3/head 2025-07-17T08:34:50.0000525Z * [new branch] gh/CaoE/3/orig -> origin/gh/CaoE/3/orig 2025-07-17T08:34:50.0006097Z * [new branch] gh/ColinPeppler/72/base -> origin/gh/ColinPeppler/72/base 2025-07-17T08:34:50.0009444Z * [new branch] gh/ColinPeppler/72/head -> origin/gh/ColinPeppler/72/head 2025-07-17T08:34:50.0012758Z * [new branch] gh/ColinPeppler/72/orig -> origin/gh/ColinPeppler/72/orig 2025-07-17T08:34:50.0022279Z * [new branch] gh/ColinPeppler/75/base -> origin/gh/ColinPeppler/75/base 2025-07-17T08:34:50.0022692Z * [new branch] gh/ColinPeppler/75/head -> origin/gh/ColinPeppler/75/head 2025-07-17T08:34:50.0024093Z * [new branch] gh/ColinPeppler/75/orig -> origin/gh/ColinPeppler/75/orig 2025-07-17T08:34:50.0028586Z * [new branch] gh/ColinPeppler/76/base -> origin/gh/ColinPeppler/76/base 2025-07-17T08:34:50.0032001Z * [new branch] gh/ColinPeppler/76/head -> origin/gh/ColinPeppler/76/head 2025-07-17T08:34:50.0036061Z * [new branch] gh/ColinPeppler/76/orig -> origin/gh/ColinPeppler/76/orig 2025-07-17T08:34:50.0041640Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-07-17T08:34:50.0045061Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-07-17T08:34:50.0049532Z * [new branch] gh/EikanWang/80/base -> origin/gh/EikanWang/80/base 2025-07-17T08:34:50.0052880Z * [new branch] gh/EikanWang/80/head -> origin/gh/EikanWang/80/head 2025-07-17T08:34:50.0056325Z * [new branch] gh/EikanWang/80/orig -> origin/gh/EikanWang/80/orig 2025-07-17T08:34:50.0060838Z * [new branch] gh/EikanWang/81/base -> origin/gh/EikanWang/81/base 2025-07-17T08:34:50.0064189Z * [new branch] gh/EikanWang/81/head -> origin/gh/EikanWang/81/head 2025-07-17T08:34:50.0067513Z * [new branch] gh/EikanWang/81/orig -> origin/gh/EikanWang/81/orig 2025-07-17T08:34:50.0073342Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-07-17T08:34:50.0076915Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-07-17T08:34:50.0082377Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-07-17T08:34:50.0085758Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-07-17T08:34:50.0089071Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-07-17T08:34:50.0093636Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-07-17T08:34:50.0097004Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-07-17T08:34:50.0100374Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-07-17T08:34:50.0104928Z * [new branch] gh/H-Huang/172/base -> origin/gh/H-Huang/172/base 2025-07-17T08:34:50.0108310Z * [new branch] gh/H-Huang/172/head -> origin/gh/H-Huang/172/head 2025-07-17T08:34:50.0111709Z * [new branch] gh/H-Huang/172/orig -> origin/gh/H-Huang/172/orig 2025-07-17T08:34:50.0116759Z * [new branch] gh/H-Huang/175/base -> origin/gh/H-Huang/175/base 2025-07-17T08:34:50.0120241Z * [new branch] gh/H-Huang/175/head -> origin/gh/H-Huang/175/head 2025-07-17T08:34:50.0123588Z * [new branch] gh/H-Huang/175/orig -> origin/gh/H-Huang/175/orig 2025-07-17T08:34:50.0127997Z * [new branch] gh/H-Huang/176/base -> origin/gh/H-Huang/176/base 2025-07-17T08:34:50.0131346Z * [new branch] gh/H-Huang/176/head -> origin/gh/H-Huang/176/head 2025-07-17T08:34:50.0134699Z * [new branch] gh/H-Huang/176/orig -> origin/gh/H-Huang/176/orig 2025-07-17T08:34:50.0139081Z * [new branch] gh/H-Huang/177/base -> origin/gh/H-Huang/177/base 2025-07-17T08:34:50.0142358Z * [new branch] gh/H-Huang/177/head -> origin/gh/H-Huang/177/head 2025-07-17T08:34:50.0145662Z * [new branch] gh/H-Huang/177/orig -> origin/gh/H-Huang/177/orig 2025-07-17T08:34:50.0150230Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-07-17T08:34:50.0153676Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-07-17T08:34:50.0156979Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-07-17T08:34:50.0161667Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-07-17T08:34:50.0165062Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-07-17T08:34:50.0168347Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-07-17T08:34:50.0172924Z * [new branch] gh/H-Huang/183/base -> origin/gh/H-Huang/183/base 2025-07-17T08:34:50.0176313Z * [new branch] gh/H-Huang/183/head -> origin/gh/H-Huang/183/head 2025-07-17T08:34:50.0179643Z * [new branch] gh/H-Huang/183/orig -> origin/gh/H-Huang/183/orig 2025-07-17T08:34:50.0184223Z * [new branch] gh/H-Huang/185/base -> origin/gh/H-Huang/185/base 2025-07-17T08:34:50.0187734Z * [new branch] gh/H-Huang/185/head -> origin/gh/H-Huang/185/head 2025-07-17T08:34:50.0191067Z * [new branch] gh/H-Huang/185/orig -> origin/gh/H-Huang/185/orig 2025-07-17T08:34:50.0195803Z * [new branch] gh/H-Huang/187/base -> origin/gh/H-Huang/187/base 2025-07-17T08:34:50.0199031Z * [new branch] gh/H-Huang/187/head -> origin/gh/H-Huang/187/head 2025-07-17T08:34:50.0202300Z * [new branch] gh/H-Huang/187/orig -> origin/gh/H-Huang/187/orig 2025-07-17T08:34:50.0207054Z * [new branch] gh/H-Huang/190/base -> origin/gh/H-Huang/190/base 2025-07-17T08:34:50.0210499Z * [new branch] gh/H-Huang/190/head -> origin/gh/H-Huang/190/head 2025-07-17T08:34:50.0213815Z * [new branch] gh/H-Huang/190/orig -> origin/gh/H-Huang/190/orig 2025-07-17T08:34:50.0218399Z * [new branch] gh/H-Huang/191/base -> origin/gh/H-Huang/191/base 2025-07-17T08:34:50.0221774Z * [new branch] gh/H-Huang/191/head -> origin/gh/H-Huang/191/head 2025-07-17T08:34:50.0225130Z * [new branch] gh/H-Huang/191/orig -> origin/gh/H-Huang/191/orig 2025-07-17T08:34:50.0229932Z * [new branch] gh/H-Huang/192/base -> origin/gh/H-Huang/192/base 2025-07-17T08:34:50.0233204Z * [new branch] gh/H-Huang/192/head -> origin/gh/H-Huang/192/head 2025-07-17T08:34:50.0236730Z * [new branch] gh/H-Huang/192/orig -> origin/gh/H-Huang/192/orig 2025-07-17T08:34:50.0241242Z * [new branch] gh/H-Huang/193/base -> origin/gh/H-Huang/193/base 2025-07-17T08:34:50.0244563Z * [new branch] gh/H-Huang/193/head -> origin/gh/H-Huang/193/head 2025-07-17T08:34:50.0248176Z * [new branch] gh/H-Huang/193/orig -> origin/gh/H-Huang/193/orig 2025-07-17T08:34:50.0252918Z * [new branch] gh/H-Huang/194/base -> origin/gh/H-Huang/194/base 2025-07-17T08:34:50.0256209Z * [new branch] gh/H-Huang/194/head -> origin/gh/H-Huang/194/head 2025-07-17T08:34:50.0259560Z * [new branch] gh/H-Huang/194/orig -> origin/gh/H-Huang/194/orig 2025-07-17T08:34:50.0265287Z * [new branch] gh/IvanKobzarev/107/base -> origin/gh/IvanKobzarev/107/base 2025-07-17T08:34:50.0268763Z * [new branch] gh/IvanKobzarev/107/head -> origin/gh/IvanKobzarev/107/head 2025-07-17T08:34:50.0272105Z * [new branch] gh/IvanKobzarev/107/orig -> origin/gh/IvanKobzarev/107/orig 2025-07-17T08:34:50.0282999Z * [new branch] gh/IvanKobzarev/109/base -> origin/gh/IvanKobzarev/109/base 2025-07-17T08:34:50.0285899Z * [new branch] gh/IvanKobzarev/109/head -> origin/gh/IvanKobzarev/109/head 2025-07-17T08:34:50.0289241Z * [new branch] gh/IvanKobzarev/109/orig -> origin/gh/IvanKobzarev/109/orig 2025-07-17T08:34:50.0294020Z * [new branch] gh/IvanKobzarev/110/base -> origin/gh/IvanKobzarev/110/base 2025-07-17T08:34:50.0297076Z * [new branch] gh/IvanKobzarev/110/head -> origin/gh/IvanKobzarev/110/head 2025-07-17T08:34:50.0300499Z * [new branch] gh/IvanKobzarev/110/orig -> origin/gh/IvanKobzarev/110/orig 2025-07-17T08:34:50.0305100Z * [new branch] gh/IvanKobzarev/111/base -> origin/gh/IvanKobzarev/111/base 2025-07-17T08:34:50.0308461Z * [new branch] gh/IvanKobzarev/111/head -> origin/gh/IvanKobzarev/111/head 2025-07-17T08:34:50.0311815Z * [new branch] gh/IvanKobzarev/111/orig -> origin/gh/IvanKobzarev/111/orig 2025-07-17T08:34:50.0316908Z * [new branch] gh/IvanKobzarev/112/base -> origin/gh/IvanKobzarev/112/base 2025-07-17T08:34:50.0320239Z * [new branch] gh/IvanKobzarev/112/head -> origin/gh/IvanKobzarev/112/head 2025-07-17T08:34:50.0323606Z * [new branch] gh/IvanKobzarev/112/orig -> origin/gh/IvanKobzarev/112/orig 2025-07-17T08:34:50.0328438Z * [new branch] gh/IvanKobzarev/114/base -> origin/gh/IvanKobzarev/114/base 2025-07-17T08:34:50.0331758Z * [new branch] gh/IvanKobzarev/114/head -> origin/gh/IvanKobzarev/114/head 2025-07-17T08:34:50.0335118Z * [new branch] gh/IvanKobzarev/114/orig -> origin/gh/IvanKobzarev/114/orig 2025-07-17T08:34:50.0339655Z * [new branch] gh/IvanKobzarev/115/base -> origin/gh/IvanKobzarev/115/base 2025-07-17T08:34:50.0342995Z * [new branch] gh/IvanKobzarev/115/head -> origin/gh/IvanKobzarev/115/head 2025-07-17T08:34:50.0346419Z * [new branch] gh/IvanKobzarev/115/orig -> origin/gh/IvanKobzarev/115/orig 2025-07-17T08:34:50.0351341Z * [new branch] gh/IvanKobzarev/116/base -> origin/gh/IvanKobzarev/116/base 2025-07-17T08:34:50.0354840Z * [new branch] gh/IvanKobzarev/116/head -> origin/gh/IvanKobzarev/116/head 2025-07-17T08:34:50.0358243Z * [new branch] gh/IvanKobzarev/116/orig -> origin/gh/IvanKobzarev/116/orig 2025-07-17T08:34:50.0362847Z * [new branch] gh/IvanKobzarev/117/base -> origin/gh/IvanKobzarev/117/base 2025-07-17T08:34:50.0366161Z * [new branch] gh/IvanKobzarev/117/head -> origin/gh/IvanKobzarev/117/head 2025-07-17T08:34:50.0369517Z * [new branch] gh/IvanKobzarev/117/orig -> origin/gh/IvanKobzarev/117/orig 2025-07-17T08:34:50.0373900Z * [new branch] gh/IvanKobzarev/118/base -> origin/gh/IvanKobzarev/118/base 2025-07-17T08:34:50.0377205Z * [new branch] gh/IvanKobzarev/118/head -> origin/gh/IvanKobzarev/118/head 2025-07-17T08:34:50.0380532Z * [new branch] gh/IvanKobzarev/118/orig -> origin/gh/IvanKobzarev/118/orig 2025-07-17T08:34:50.0385098Z * [new branch] gh/IvanKobzarev/119/base -> origin/gh/IvanKobzarev/119/base 2025-07-17T08:34:50.0388462Z * [new branch] gh/IvanKobzarev/119/head -> origin/gh/IvanKobzarev/119/head 2025-07-17T08:34:50.0392009Z * [new branch] gh/IvanKobzarev/119/orig -> origin/gh/IvanKobzarev/119/orig 2025-07-17T08:34:50.0399228Z * [new branch] gh/IvanKobzarev/120/base -> origin/gh/IvanKobzarev/120/base 2025-07-17T08:34:50.0402462Z * [new branch] gh/IvanKobzarev/120/head -> origin/gh/IvanKobzarev/120/head 2025-07-17T08:34:50.0405786Z * [new branch] gh/IvanKobzarev/120/orig -> origin/gh/IvanKobzarev/120/orig 2025-07-17T08:34:50.0410442Z * [new branch] gh/IvanKobzarev/121/base -> origin/gh/IvanKobzarev/121/base 2025-07-17T08:34:50.0413835Z * [new branch] gh/IvanKobzarev/121/head -> origin/gh/IvanKobzarev/121/head 2025-07-17T08:34:50.0417178Z * [new branch] gh/IvanKobzarev/121/orig -> origin/gh/IvanKobzarev/121/orig 2025-07-17T08:34:50.0421766Z * [new branch] gh/IvanKobzarev/122/base -> origin/gh/IvanKobzarev/122/base 2025-07-17T08:34:50.0425214Z * [new branch] gh/IvanKobzarev/122/head -> origin/gh/IvanKobzarev/122/head 2025-07-17T08:34:50.0428341Z * [new branch] gh/IvanKobzarev/122/orig -> origin/gh/IvanKobzarev/122/orig 2025-07-17T08:34:50.0433029Z * [new branch] gh/IvanKobzarev/123/base -> origin/gh/IvanKobzarev/123/base 2025-07-17T08:34:50.0437586Z * [new branch] gh/IvanKobzarev/123/head -> origin/gh/IvanKobzarev/123/head 2025-07-17T08:34:50.0440023Z * [new branch] gh/IvanKobzarev/123/orig -> origin/gh/IvanKobzarev/123/orig 2025-07-17T08:34:50.0444777Z * [new branch] gh/IvanKobzarev/124/base -> origin/gh/IvanKobzarev/124/base 2025-07-17T08:34:50.0448102Z * [new branch] gh/IvanKobzarev/124/head -> origin/gh/IvanKobzarev/124/head 2025-07-17T08:34:50.0451490Z * [new branch] gh/IvanKobzarev/124/orig -> origin/gh/IvanKobzarev/124/orig 2025-07-17T08:34:50.0456045Z * [new branch] gh/IvanKobzarev/125/base -> origin/gh/IvanKobzarev/125/base 2025-07-17T08:34:50.0459348Z * [new branch] gh/IvanKobzarev/125/head -> origin/gh/IvanKobzarev/125/head 2025-07-17T08:34:50.0462714Z * [new branch] gh/IvanKobzarev/125/orig -> origin/gh/IvanKobzarev/125/orig 2025-07-17T08:34:50.0467357Z * [new branch] gh/IvanKobzarev/126/base -> origin/gh/IvanKobzarev/126/base 2025-07-17T08:34:50.0470726Z * [new branch] gh/IvanKobzarev/126/head -> origin/gh/IvanKobzarev/126/head 2025-07-17T08:34:50.0474131Z * [new branch] gh/IvanKobzarev/126/orig -> origin/gh/IvanKobzarev/126/orig 2025-07-17T08:34:50.0478939Z * [new branch] gh/IvanKobzarev/127/base -> origin/gh/IvanKobzarev/127/base 2025-07-17T08:34:50.0482180Z * [new branch] gh/IvanKobzarev/127/head -> origin/gh/IvanKobzarev/127/head 2025-07-17T08:34:50.0485622Z * [new branch] gh/IvanKobzarev/127/orig -> origin/gh/IvanKobzarev/127/orig 2025-07-17T08:34:50.0490233Z * [new branch] gh/IvanKobzarev/128/base -> origin/gh/IvanKobzarev/128/base 2025-07-17T08:34:50.0493651Z * [new branch] gh/IvanKobzarev/128/head -> origin/gh/IvanKobzarev/128/head 2025-07-17T08:34:50.0496939Z * [new branch] gh/IvanKobzarev/128/orig -> origin/gh/IvanKobzarev/128/orig 2025-07-17T08:34:50.0502527Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-07-17T08:34:50.0505937Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-07-17T08:34:50.0510880Z * [new branch] gh/NikhilAPatel/16/base -> origin/gh/NikhilAPatel/16/base 2025-07-17T08:34:50.0514362Z * [new branch] gh/NikhilAPatel/16/head -> origin/gh/NikhilAPatel/16/head 2025-07-17T08:34:50.0517703Z * [new branch] gh/NikhilAPatel/16/orig -> origin/gh/NikhilAPatel/16/orig 2025-07-17T08:34:50.0522280Z * [new branch] gh/NikhilAPatel/17/base -> origin/gh/NikhilAPatel/17/base 2025-07-17T08:34:50.0525647Z * [new branch] gh/NikhilAPatel/17/head -> origin/gh/NikhilAPatel/17/head 2025-07-17T08:34:50.0529083Z * [new branch] gh/NikhilAPatel/17/orig -> origin/gh/NikhilAPatel/17/orig 2025-07-17T08:34:50.0533595Z * [new branch] gh/NikhilAPatel/18/base -> origin/gh/NikhilAPatel/18/base 2025-07-17T08:34:50.0537165Z * [new branch] gh/NikhilAPatel/18/head -> origin/gh/NikhilAPatel/18/head 2025-07-17T08:34:50.0540503Z * [new branch] gh/NikhilAPatel/18/orig -> origin/gh/NikhilAPatel/18/orig 2025-07-17T08:34:50.0544913Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-07-17T08:34:50.0548188Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-07-17T08:34:50.0552875Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-07-17T08:34:50.0556760Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-07-17T08:34:50.0561207Z * [new branch] gh/NikhilAPatel/8/base -> origin/gh/NikhilAPatel/8/base 2025-07-17T08:34:50.0564488Z * [new branch] gh/NikhilAPatel/8/head -> origin/gh/NikhilAPatel/8/head 2025-07-17T08:34:50.0567923Z * [new branch] gh/NikhilAPatel/8/orig -> origin/gh/NikhilAPatel/8/orig 2025-07-17T08:34:50.0572848Z * [new branch] gh/NikhilAPatel/9/base -> origin/gh/NikhilAPatel/9/base 2025-07-17T08:34:50.0576113Z * [new branch] gh/NikhilAPatel/9/head -> origin/gh/NikhilAPatel/9/head 2025-07-17T08:34:50.0579490Z * [new branch] gh/NikhilAPatel/9/orig -> origin/gh/NikhilAPatel/9/orig 2025-07-17T08:34:50.0584936Z * [new branch] gh/PaliC/1/base -> origin/gh/PaliC/1/base 2025-07-17T08:34:50.0588183Z * [new branch] gh/PaliC/1/head -> origin/gh/PaliC/1/head 2025-07-17T08:34:50.0591528Z * [new branch] gh/PaliC/1/orig -> origin/gh/PaliC/1/orig 2025-07-17T08:34:50.0596155Z * [new branch] gh/PaliC/10/base -> origin/gh/PaliC/10/base 2025-07-17T08:34:50.0599473Z * [new branch] gh/PaliC/10/head -> origin/gh/PaliC/10/head 2025-07-17T08:34:50.0603849Z * [new branch] gh/PaliC/11/base -> origin/gh/PaliC/11/base 2025-07-17T08:34:50.0607187Z * [new branch] gh/PaliC/11/head -> origin/gh/PaliC/11/head 2025-07-17T08:34:50.0611823Z * [new branch] gh/PaliC/12/base -> origin/gh/PaliC/12/base 2025-07-17T08:34:50.0615328Z * [new branch] gh/PaliC/12/head -> origin/gh/PaliC/12/head 2025-07-17T08:34:50.0618731Z * [new branch] gh/PaliC/12/orig -> origin/gh/PaliC/12/orig 2025-07-17T08:34:50.0623160Z * [new branch] gh/PaliC/13/base -> origin/gh/PaliC/13/base 2025-07-17T08:34:50.0626522Z * [new branch] gh/PaliC/13/head -> origin/gh/PaliC/13/head 2025-07-17T08:34:50.0629850Z * [new branch] gh/PaliC/13/orig -> origin/gh/PaliC/13/orig 2025-07-17T08:34:50.0636544Z * [new branch] gh/PaliC/14/base -> origin/gh/PaliC/14/base 2025-07-17T08:34:50.0639973Z * [new branch] gh/PaliC/14/head -> origin/gh/PaliC/14/head 2025-07-17T08:34:50.0643414Z * [new branch] gh/PaliC/14/orig -> origin/gh/PaliC/14/orig 2025-07-17T08:34:50.0647900Z * [new branch] gh/PaliC/15/base -> origin/gh/PaliC/15/base 2025-07-17T08:34:50.0651164Z * [new branch] gh/PaliC/15/head -> origin/gh/PaliC/15/head 2025-07-17T08:34:50.0654463Z * [new branch] gh/PaliC/15/orig -> origin/gh/PaliC/15/orig 2025-07-17T08:34:50.0659155Z * [new branch] gh/PaliC/16/base -> origin/gh/PaliC/16/base 2025-07-17T08:34:50.0662691Z * [new branch] gh/PaliC/16/head -> origin/gh/PaliC/16/head 2025-07-17T08:34:50.0666032Z * [new branch] gh/PaliC/16/orig -> origin/gh/PaliC/16/orig 2025-07-17T08:34:50.0670544Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-07-17T08:34:50.0673866Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-07-17T08:34:50.0677419Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-07-17T08:34:50.0681770Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-07-17T08:34:50.0685108Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-07-17T08:34:50.0688454Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-07-17T08:34:50.0693037Z * [new branch] gh/PaliC/19/base -> origin/gh/PaliC/19/base 2025-07-17T08:34:50.0696740Z * [new branch] gh/PaliC/19/head -> origin/gh/PaliC/19/head 2025-07-17T08:34:50.0699864Z * [new branch] gh/PaliC/19/orig -> origin/gh/PaliC/19/orig 2025-07-17T08:34:50.0704601Z * [new branch] gh/PaliC/2/base -> origin/gh/PaliC/2/base 2025-07-17T08:34:50.0708049Z * [new branch] gh/PaliC/2/head -> origin/gh/PaliC/2/head 2025-07-17T08:34:50.0711389Z * [new branch] gh/PaliC/2/orig -> origin/gh/PaliC/2/orig 2025-07-17T08:34:50.0716077Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-07-17T08:34:50.0719375Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-07-17T08:34:50.0722598Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-07-17T08:34:50.0727044Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-07-17T08:34:50.0730376Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-07-17T08:34:50.0733704Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-07-17T08:34:50.0738090Z * [new branch] gh/PaliC/22/base -> origin/gh/PaliC/22/base 2025-07-17T08:34:50.0741387Z * [new branch] gh/PaliC/22/head -> origin/gh/PaliC/22/head 2025-07-17T08:34:50.0744697Z * [new branch] gh/PaliC/22/orig -> origin/gh/PaliC/22/orig 2025-07-17T08:34:50.0749495Z * [new branch] gh/PaliC/3/base -> origin/gh/PaliC/3/base 2025-07-17T08:34:50.0752943Z * [new branch] gh/PaliC/3/head -> origin/gh/PaliC/3/head 2025-07-17T08:34:50.0757685Z * [new branch] gh/PaliC/4/base -> origin/gh/PaliC/4/base 2025-07-17T08:34:50.0760759Z * [new branch] gh/PaliC/4/head -> origin/gh/PaliC/4/head 2025-07-17T08:34:50.0765159Z * [new branch] gh/PaliC/5/base -> origin/gh/PaliC/5/base 2025-07-17T08:34:50.0768482Z * [new branch] gh/PaliC/5/head -> origin/gh/PaliC/5/head 2025-07-17T08:34:50.0772798Z * [new branch] gh/PaliC/6/base -> origin/gh/PaliC/6/base 2025-07-17T08:34:50.0776098Z * [new branch] gh/PaliC/6/head -> origin/gh/PaliC/6/head 2025-07-17T08:34:50.0780435Z * [new branch] gh/PaliC/7/base -> origin/gh/PaliC/7/base 2025-07-17T08:34:50.0783780Z * [new branch] gh/PaliC/7/head -> origin/gh/PaliC/7/head 2025-07-17T08:34:50.0788231Z * [new branch] gh/PaliC/8/base -> origin/gh/PaliC/8/base 2025-07-17T08:34:50.0791579Z * [new branch] gh/PaliC/8/head -> origin/gh/PaliC/8/head 2025-07-17T08:34:50.0796225Z * [new branch] gh/PaliC/9/base -> origin/gh/PaliC/9/base 2025-07-17T08:34:50.0799490Z * [new branch] gh/PaliC/9/head -> origin/gh/PaliC/9/head 2025-07-17T08:34:50.0805057Z * [new branch] gh/PaulZhang12/1/base -> origin/gh/PaulZhang12/1/base 2025-07-17T08:34:50.0808374Z * [new branch] gh/PaulZhang12/1/head -> origin/gh/PaulZhang12/1/head 2025-07-17T08:34:50.0811869Z * [new branch] gh/PaulZhang12/1/orig -> origin/gh/PaulZhang12/1/orig 2025-07-17T08:34:50.0816394Z * [new branch] gh/PaulZhang12/17/base -> origin/gh/PaulZhang12/17/base 2025-07-17T08:34:50.0819712Z * [new branch] gh/PaulZhang12/17/head -> origin/gh/PaulZhang12/17/head 2025-07-17T08:34:50.0824211Z * [new branch] gh/PaulZhang12/4/base -> origin/gh/PaulZhang12/4/base 2025-07-17T08:34:50.0827568Z * [new branch] gh/PaulZhang12/4/head -> origin/gh/PaulZhang12/4/head 2025-07-17T08:34:50.0831255Z * [new branch] gh/PaulZhang12/4/orig -> origin/gh/PaulZhang12/4/orig 2025-07-17T08:34:50.0837263Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-07-17T08:34:50.0840368Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-07-17T08:34:50.0845967Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-07-17T08:34:50.0850424Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-07-17T08:34:50.0854778Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-07-17T08:34:50.0859351Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-07-17T08:34:50.0863713Z * [new branch] gh/Sidharth123-cpu/28/base -> origin/gh/Sidharth123-cpu/28/base 2025-07-17T08:34:50.0866999Z * [new branch] gh/Sidharth123-cpu/28/head -> origin/gh/Sidharth123-cpu/28/head 2025-07-17T08:34:50.0870303Z * [new branch] gh/Sidharth123-cpu/28/orig -> origin/gh/Sidharth123-cpu/28/orig 2025-07-17T08:34:50.0876285Z * [new branch] gh/Sidharth123-cpu/32/base -> origin/gh/Sidharth123-cpu/32/base 2025-07-17T08:34:50.0879567Z * [new branch] gh/Sidharth123-cpu/32/head -> origin/gh/Sidharth123-cpu/32/head 2025-07-17T08:34:50.0882929Z * [new branch] gh/Sidharth123-cpu/32/orig -> origin/gh/Sidharth123-cpu/32/orig 2025-07-17T08:34:50.0887585Z * [new branch] gh/Sidharth123-cpu/33/base -> origin/gh/Sidharth123-cpu/33/base 2025-07-17T08:34:50.0890992Z * [new branch] gh/Sidharth123-cpu/33/head -> origin/gh/Sidharth123-cpu/33/head 2025-07-17T08:34:50.0894349Z * [new branch] gh/Sidharth123-cpu/33/orig -> origin/gh/Sidharth123-cpu/33/orig 2025-07-17T08:34:50.0899078Z * [new branch] gh/Sidharth123-cpu/34/base -> origin/gh/Sidharth123-cpu/34/base 2025-07-17T08:34:50.0902605Z * [new branch] gh/Sidharth123-cpu/34/head -> origin/gh/Sidharth123-cpu/34/head 2025-07-17T08:34:50.0906004Z * [new branch] gh/Sidharth123-cpu/34/orig -> origin/gh/Sidharth123-cpu/34/orig 2025-07-17T08:34:50.0910686Z * [new branch] gh/Sidharth123-cpu/35/base -> origin/gh/Sidharth123-cpu/35/base 2025-07-17T08:34:50.0914215Z * [new branch] gh/Sidharth123-cpu/35/head -> origin/gh/Sidharth123-cpu/35/head 2025-07-17T08:34:50.0917659Z * [new branch] gh/Sidharth123-cpu/35/orig -> origin/gh/Sidharth123-cpu/35/orig 2025-07-17T08:34:50.0922486Z * [new branch] gh/Sidharth123-cpu/36/base -> origin/gh/Sidharth123-cpu/36/base 2025-07-17T08:34:50.0925915Z * [new branch] gh/Sidharth123-cpu/36/head -> origin/gh/Sidharth123-cpu/36/head 2025-07-17T08:34:50.0929257Z * [new branch] gh/Sidharth123-cpu/36/orig -> origin/gh/Sidharth123-cpu/36/orig 2025-07-17T08:34:50.0933797Z * [new branch] gh/Sidharth123-cpu/37/base -> origin/gh/Sidharth123-cpu/37/base 2025-07-17T08:34:50.0937153Z * [new branch] gh/Sidharth123-cpu/37/head -> origin/gh/Sidharth123-cpu/37/head 2025-07-17T08:34:50.0940601Z * [new branch] gh/Sidharth123-cpu/37/orig -> origin/gh/Sidharth123-cpu/37/orig 2025-07-17T08:34:50.0945150Z * [new branch] gh/Sidharth123-cpu/38/base -> origin/gh/Sidharth123-cpu/38/base 2025-07-17T08:34:50.0948474Z * [new branch] gh/Sidharth123-cpu/38/head -> origin/gh/Sidharth123-cpu/38/head 2025-07-17T08:34:50.0951802Z * [new branch] gh/Sidharth123-cpu/38/orig -> origin/gh/Sidharth123-cpu/38/orig 2025-07-17T08:34:50.0956534Z * [new branch] gh/Sidharth123-cpu/39/base -> origin/gh/Sidharth123-cpu/39/base 2025-07-17T08:34:50.0959819Z * [new branch] gh/Sidharth123-cpu/39/head -> origin/gh/Sidharth123-cpu/39/head 2025-07-17T08:34:50.0963160Z * [new branch] gh/Sidharth123-cpu/39/orig -> origin/gh/Sidharth123-cpu/39/orig 2025-07-17T08:34:50.0967874Z * [new branch] gh/Sidharth123-cpu/40/base -> origin/gh/Sidharth123-cpu/40/base 2025-07-17T08:34:50.0971026Z * [new branch] gh/Sidharth123-cpu/40/head -> origin/gh/Sidharth123-cpu/40/head 2025-07-17T08:34:50.0974351Z * [new branch] gh/Sidharth123-cpu/40/orig -> origin/gh/Sidharth123-cpu/40/orig 2025-07-17T08:34:50.0978984Z * [new branch] gh/Sidharth123-cpu/41/base -> origin/gh/Sidharth123-cpu/41/base 2025-07-17T08:34:50.0982261Z * [new branch] gh/Sidharth123-cpu/41/head -> origin/gh/Sidharth123-cpu/41/head 2025-07-17T08:34:50.0985625Z * [new branch] gh/Sidharth123-cpu/41/orig -> origin/gh/Sidharth123-cpu/41/orig 2025-07-17T08:34:50.0990428Z * [new branch] gh/Sidharth123-cpu/42/base -> origin/gh/Sidharth123-cpu/42/base 2025-07-17T08:34:50.0993827Z * [new branch] gh/Sidharth123-cpu/42/head -> origin/gh/Sidharth123-cpu/42/head 2025-07-17T08:34:50.0997319Z * [new branch] gh/Sidharth123-cpu/42/orig -> origin/gh/Sidharth123-cpu/42/orig 2025-07-17T08:34:50.1002050Z * [new branch] gh/Sidharth123-cpu/43/base -> origin/gh/Sidharth123-cpu/43/base 2025-07-17T08:34:50.1005311Z * [new branch] gh/Sidharth123-cpu/43/head -> origin/gh/Sidharth123-cpu/43/head 2025-07-17T08:34:50.1008649Z * [new branch] gh/Sidharth123-cpu/43/orig -> origin/gh/Sidharth123-cpu/43/orig 2025-07-17T08:34:50.1014233Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-07-17T08:34:50.1017484Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-07-17T08:34:50.1022049Z * [new branch] gh/StrongerXi/103/base -> origin/gh/StrongerXi/103/base 2025-07-17T08:34:50.1025377Z * [new branch] gh/StrongerXi/103/head -> origin/gh/StrongerXi/103/head 2025-07-17T08:34:50.1028756Z * [new branch] gh/StrongerXi/103/orig -> origin/gh/StrongerXi/103/orig 2025-07-17T08:34:50.1033274Z * [new branch] gh/StrongerXi/126/base -> origin/gh/StrongerXi/126/base 2025-07-17T08:34:50.1036776Z * [new branch] gh/StrongerXi/126/head -> origin/gh/StrongerXi/126/head 2025-07-17T08:34:50.1040103Z * [new branch] gh/StrongerXi/126/orig -> origin/gh/StrongerXi/126/orig 2025-07-17T08:34:50.1044607Z * [new branch] gh/StrongerXi/127/base -> origin/gh/StrongerXi/127/base 2025-07-17T08:34:50.1047965Z * [new branch] gh/StrongerXi/127/head -> origin/gh/StrongerXi/127/head 2025-07-17T08:34:50.1051386Z * [new branch] gh/StrongerXi/127/orig -> origin/gh/StrongerXi/127/orig 2025-07-17T08:34:50.1055970Z * [new branch] gh/StrongerXi/128/base -> origin/gh/StrongerXi/128/base 2025-07-17T08:34:50.1059311Z * [new branch] gh/StrongerXi/128/head -> origin/gh/StrongerXi/128/head 2025-07-17T08:34:50.1062639Z * [new branch] gh/StrongerXi/128/orig -> origin/gh/StrongerXi/128/orig 2025-07-17T08:34:50.1067304Z * [new branch] gh/StrongerXi/129/base -> origin/gh/StrongerXi/129/base 2025-07-17T08:34:50.1070585Z * [new branch] gh/StrongerXi/129/head -> origin/gh/StrongerXi/129/head 2025-07-17T08:34:50.1073996Z * [new branch] gh/StrongerXi/129/orig -> origin/gh/StrongerXi/129/orig 2025-07-17T08:34:50.1078654Z * [new branch] gh/StrongerXi/130/base -> origin/gh/StrongerXi/130/base 2025-07-17T08:34:50.1082004Z * [new branch] gh/StrongerXi/130/head -> origin/gh/StrongerXi/130/head 2025-07-17T08:34:50.1085352Z * [new branch] gh/StrongerXi/130/orig -> origin/gh/StrongerXi/130/orig 2025-07-17T08:34:50.1090038Z * [new branch] gh/StrongerXi/131/base -> origin/gh/StrongerXi/131/base 2025-07-17T08:34:50.1093586Z * [new branch] gh/StrongerXi/131/head -> origin/gh/StrongerXi/131/head 2025-07-17T08:34:50.1097179Z * [new branch] gh/StrongerXi/131/orig -> origin/gh/StrongerXi/131/orig 2025-07-17T08:34:50.1101629Z * [new branch] gh/StrongerXi/132/base -> origin/gh/StrongerXi/132/base 2025-07-17T08:34:50.1104954Z * [new branch] gh/StrongerXi/132/head -> origin/gh/StrongerXi/132/head 2025-07-17T08:34:50.1108287Z * [new branch] gh/StrongerXi/132/orig -> origin/gh/StrongerXi/132/orig 2025-07-17T08:34:50.1113098Z * [new branch] gh/StrongerXi/133/base -> origin/gh/StrongerXi/133/base 2025-07-17T08:34:50.1119292Z * [new branch] gh/StrongerXi/133/head -> origin/gh/StrongerXi/133/head 2025-07-17T08:34:50.1122840Z * [new branch] gh/StrongerXi/133/orig -> origin/gh/StrongerXi/133/orig 2025-07-17T08:34:50.1127363Z * [new branch] gh/StrongerXi/134/base -> origin/gh/StrongerXi/134/base 2025-07-17T08:34:50.1130709Z * [new branch] gh/StrongerXi/134/head -> origin/gh/StrongerXi/134/head 2025-07-17T08:34:50.1133982Z * [new branch] gh/StrongerXi/134/orig -> origin/gh/StrongerXi/134/orig 2025-07-17T08:34:50.1138586Z * [new branch] gh/StrongerXi/135/base -> origin/gh/StrongerXi/135/base 2025-07-17T08:34:50.1142009Z * [new branch] gh/StrongerXi/135/head -> origin/gh/StrongerXi/135/head 2025-07-17T08:34:50.1145389Z * [new branch] gh/StrongerXi/135/orig -> origin/gh/StrongerXi/135/orig 2025-07-17T08:34:50.1149944Z * [new branch] gh/StrongerXi/136/base -> origin/gh/StrongerXi/136/base 2025-07-17T08:34:50.1153298Z * [new branch] gh/StrongerXi/136/head -> origin/gh/StrongerXi/136/head 2025-07-17T08:34:50.1156982Z * [new branch] gh/StrongerXi/136/orig -> origin/gh/StrongerXi/136/orig 2025-07-17T08:34:50.1161501Z * [new branch] gh/StrongerXi/137/base -> origin/gh/StrongerXi/137/base 2025-07-17T08:34:50.1164792Z * [new branch] gh/StrongerXi/137/head -> origin/gh/StrongerXi/137/head 2025-07-17T08:34:50.1168326Z * [new branch] gh/StrongerXi/137/orig -> origin/gh/StrongerXi/137/orig 2025-07-17T08:34:50.1172732Z * [new branch] gh/StrongerXi/138/base -> origin/gh/StrongerXi/138/base 2025-07-17T08:34:50.1176085Z * [new branch] gh/StrongerXi/138/head -> origin/gh/StrongerXi/138/head 2025-07-17T08:34:50.1179394Z * [new branch] gh/StrongerXi/138/orig -> origin/gh/StrongerXi/138/orig 2025-07-17T08:34:50.1184105Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-07-17T08:34:50.1187450Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-07-17T08:34:50.1191962Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-07-17T08:34:50.1195445Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-07-17T08:34:50.1200852Z * [new branch] gh/Xia-Weiwen/40/base -> origin/gh/Xia-Weiwen/40/base 2025-07-17T08:34:50.1204209Z * [new branch] gh/Xia-Weiwen/40/head -> origin/gh/Xia-Weiwen/40/head 2025-07-17T08:34:50.1207485Z * [new branch] gh/Xia-Weiwen/40/orig -> origin/gh/Xia-Weiwen/40/orig 2025-07-17T08:34:50.1213242Z * [new branch] gh/XilunWu/110/base -> origin/gh/XilunWu/110/base 2025-07-17T08:34:50.1216712Z * [new branch] gh/XilunWu/110/head -> origin/gh/XilunWu/110/head 2025-07-17T08:34:50.1220046Z * [new branch] gh/XilunWu/110/orig -> origin/gh/XilunWu/110/orig 2025-07-17T08:34:50.1224560Z * [new branch] gh/XilunWu/115/base -> origin/gh/XilunWu/115/base 2025-07-17T08:34:50.1227923Z * [new branch] gh/XilunWu/115/head -> origin/gh/XilunWu/115/head 2025-07-17T08:34:50.1231308Z * [new branch] gh/XilunWu/115/orig -> origin/gh/XilunWu/115/orig 2025-07-17T08:34:50.1236878Z * [new branch] gh/XilunWu/116/base -> origin/gh/XilunWu/116/base 2025-07-17T08:34:50.1240608Z * [new branch] gh/XilunWu/116/head -> origin/gh/XilunWu/116/head 2025-07-17T08:34:50.1244071Z * [new branch] gh/XilunWu/116/orig -> origin/gh/XilunWu/116/orig 2025-07-17T08:34:50.1248937Z * [new branch] gh/XilunWu/117/base -> origin/gh/XilunWu/117/base 2025-07-17T08:34:50.1252750Z * [new branch] gh/XilunWu/117/head -> origin/gh/XilunWu/117/head 2025-07-17T08:34:50.1256261Z * [new branch] gh/XilunWu/117/orig -> origin/gh/XilunWu/117/orig 2025-07-17T08:34:50.1260985Z * [new branch] gh/XilunWu/118/base -> origin/gh/XilunWu/118/base 2025-07-17T08:34:50.1264472Z * [new branch] gh/XilunWu/118/head -> origin/gh/XilunWu/118/head 2025-07-17T08:34:50.1267765Z * [new branch] gh/XilunWu/118/orig -> origin/gh/XilunWu/118/orig 2025-07-17T08:34:50.1272403Z * [new branch] gh/XilunWu/119/base -> origin/gh/XilunWu/119/base 2025-07-17T08:34:50.1276165Z * [new branch] gh/XilunWu/119/head -> origin/gh/XilunWu/119/head 2025-07-17T08:34:50.1280322Z * [new branch] gh/XilunWu/119/orig -> origin/gh/XilunWu/119/orig 2025-07-17T08:34:50.1284018Z * [new branch] gh/XilunWu/126/base -> origin/gh/XilunWu/126/base 2025-07-17T08:34:50.1287290Z * [new branch] gh/XilunWu/126/head -> origin/gh/XilunWu/126/head 2025-07-17T08:34:50.1290847Z * [new branch] gh/XilunWu/126/orig -> origin/gh/XilunWu/126/orig 2025-07-17T08:34:50.1295591Z * [new branch] gh/XilunWu/127/base -> origin/gh/XilunWu/127/base 2025-07-17T08:34:50.1298931Z * [new branch] gh/XilunWu/127/head -> origin/gh/XilunWu/127/head 2025-07-17T08:34:50.1302441Z * [new branch] gh/XilunWu/127/orig -> origin/gh/XilunWu/127/orig 2025-07-17T08:34:50.1307056Z * [new branch] gh/XilunWu/129/base -> origin/gh/XilunWu/129/base 2025-07-17T08:34:50.1310363Z * [new branch] gh/XilunWu/129/head -> origin/gh/XilunWu/129/head 2025-07-17T08:34:50.1313676Z * [new branch] gh/XilunWu/129/orig -> origin/gh/XilunWu/129/orig 2025-07-17T08:34:50.1318578Z * [new branch] gh/XilunWu/130/base -> origin/gh/XilunWu/130/base 2025-07-17T08:34:50.1321786Z * [new branch] gh/XilunWu/130/head -> origin/gh/XilunWu/130/head 2025-07-17T08:34:50.1325169Z * [new branch] gh/XilunWu/130/orig -> origin/gh/XilunWu/130/orig 2025-07-17T08:34:50.1329802Z * [new branch] gh/XilunWu/131/base -> origin/gh/XilunWu/131/base 2025-07-17T08:34:50.1333351Z * [new branch] gh/XilunWu/131/head -> origin/gh/XilunWu/131/head 2025-07-17T08:34:50.1336768Z * [new branch] gh/XilunWu/131/orig -> origin/gh/XilunWu/131/orig 2025-07-17T08:34:50.1341424Z * [new branch] gh/XilunWu/133/base -> origin/gh/XilunWu/133/base 2025-07-17T08:34:50.1344670Z * [new branch] gh/XilunWu/133/head -> origin/gh/XilunWu/133/head 2025-07-17T08:34:50.1348216Z * [new branch] gh/XilunWu/133/orig -> origin/gh/XilunWu/133/orig 2025-07-17T08:34:50.1352827Z * [new branch] gh/XilunWu/136/base -> origin/gh/XilunWu/136/base 2025-07-17T08:34:50.1356373Z * [new branch] gh/XilunWu/136/head -> origin/gh/XilunWu/136/head 2025-07-17T08:34:50.1359693Z * [new branch] gh/XilunWu/136/orig -> origin/gh/XilunWu/136/orig 2025-07-17T08:34:50.1364309Z * [new branch] gh/XilunWu/139/base -> origin/gh/XilunWu/139/base 2025-07-17T08:34:50.1367623Z * [new branch] gh/XilunWu/139/head -> origin/gh/XilunWu/139/head 2025-07-17T08:34:50.1371061Z * [new branch] gh/XilunWu/139/orig -> origin/gh/XilunWu/139/orig 2025-07-17T08:34:50.1375494Z * [new branch] gh/XilunWu/143/base -> origin/gh/XilunWu/143/base 2025-07-17T08:34:50.1378943Z * [new branch] gh/XilunWu/143/head -> origin/gh/XilunWu/143/head 2025-07-17T08:34:50.1382249Z * [new branch] gh/XilunWu/143/orig -> origin/gh/XilunWu/143/orig 2025-07-17T08:34:50.1387035Z * [new branch] gh/XilunWu/144/base -> origin/gh/XilunWu/144/base 2025-07-17T08:34:50.1390408Z * [new branch] gh/XilunWu/144/head -> origin/gh/XilunWu/144/head 2025-07-17T08:34:50.1393831Z * [new branch] gh/XilunWu/144/orig -> origin/gh/XilunWu/144/orig 2025-07-17T08:34:50.1398702Z * [new branch] gh/XilunWu/145/base -> origin/gh/XilunWu/145/base 2025-07-17T08:34:50.1402213Z * [new branch] gh/XilunWu/145/head -> origin/gh/XilunWu/145/head 2025-07-17T08:34:50.1405631Z * [new branch] gh/XilunWu/145/orig -> origin/gh/XilunWu/145/orig 2025-07-17T08:34:50.1410040Z * [new branch] gh/XilunWu/146/base -> origin/gh/XilunWu/146/base 2025-07-17T08:34:50.1413366Z * [new branch] gh/XilunWu/146/head -> origin/gh/XilunWu/146/head 2025-07-17T08:34:50.1416680Z * [new branch] gh/XilunWu/146/orig -> origin/gh/XilunWu/146/orig 2025-07-17T08:34:50.1421216Z * [new branch] gh/XilunWu/147/base -> origin/gh/XilunWu/147/base 2025-07-17T08:34:50.1424541Z * [new branch] gh/XilunWu/147/head -> origin/gh/XilunWu/147/head 2025-07-17T08:34:50.1427832Z * [new branch] gh/XilunWu/147/orig -> origin/gh/XilunWu/147/orig 2025-07-17T08:34:50.1432246Z * [new branch] gh/XilunWu/148/base -> origin/gh/XilunWu/148/base 2025-07-17T08:34:50.1435818Z * [new branch] gh/XilunWu/148/head -> origin/gh/XilunWu/148/head 2025-07-17T08:34:50.1439212Z * [new branch] gh/XilunWu/148/orig -> origin/gh/XilunWu/148/orig 2025-07-17T08:34:50.1443610Z * [new branch] gh/XilunWu/149/base -> origin/gh/XilunWu/149/base 2025-07-17T08:34:50.1446941Z * [new branch] gh/XilunWu/149/head -> origin/gh/XilunWu/149/head 2025-07-17T08:34:50.1450307Z * [new branch] gh/XilunWu/149/orig -> origin/gh/XilunWu/149/orig 2025-07-17T08:34:50.1454702Z * [new branch] gh/XilunWu/150/base -> origin/gh/XilunWu/150/base 2025-07-17T08:34:50.1458047Z * [new branch] gh/XilunWu/150/head -> origin/gh/XilunWu/150/head 2025-07-17T08:34:50.1461371Z * [new branch] gh/XilunWu/150/orig -> origin/gh/XilunWu/150/orig 2025-07-17T08:34:50.1465999Z * [new branch] gh/XilunWu/151/base -> origin/gh/XilunWu/151/base 2025-07-17T08:34:50.1469331Z * [new branch] gh/XilunWu/151/head -> origin/gh/XilunWu/151/head 2025-07-17T08:34:50.1472664Z * [new branch] gh/XilunWu/151/orig -> origin/gh/XilunWu/151/orig 2025-07-17T08:34:50.1477534Z * [new branch] gh/XilunWu/152/base -> origin/gh/XilunWu/152/base 2025-07-17T08:34:50.1480976Z * [new branch] gh/XilunWu/152/head -> origin/gh/XilunWu/152/head 2025-07-17T08:34:50.1484460Z * [new branch] gh/XilunWu/152/orig -> origin/gh/XilunWu/152/orig 2025-07-17T08:34:50.1488892Z * [new branch] gh/XilunWu/153/base -> origin/gh/XilunWu/153/base 2025-07-17T08:34:50.1492253Z * [new branch] gh/XilunWu/153/head -> origin/gh/XilunWu/153/head 2025-07-17T08:34:50.1495557Z * [new branch] gh/XilunWu/153/orig -> origin/gh/XilunWu/153/orig 2025-07-17T08:34:50.1499952Z * [new branch] gh/XilunWu/154/base -> origin/gh/XilunWu/154/base 2025-07-17T08:34:50.1503346Z * [new branch] gh/XilunWu/154/head -> origin/gh/XilunWu/154/head 2025-07-17T08:34:50.1506947Z * [new branch] gh/XilunWu/154/orig -> origin/gh/XilunWu/154/orig 2025-07-17T08:34:50.1511462Z * [new branch] gh/XilunWu/155/base -> origin/gh/XilunWu/155/base 2025-07-17T08:34:50.1515131Z * [new branch] gh/XilunWu/155/head -> origin/gh/XilunWu/155/head 2025-07-17T08:34:50.1518392Z * [new branch] gh/XilunWu/155/orig -> origin/gh/XilunWu/155/orig 2025-07-17T08:34:50.1523081Z * [new branch] gh/XilunWu/156/base -> origin/gh/XilunWu/156/base 2025-07-17T08:34:50.1526472Z * [new branch] gh/XilunWu/156/head -> origin/gh/XilunWu/156/head 2025-07-17T08:34:50.1529993Z * [new branch] gh/XilunWu/156/orig -> origin/gh/XilunWu/156/orig 2025-07-17T08:34:50.1534476Z * [new branch] gh/XilunWu/157/base -> origin/gh/XilunWu/157/base 2025-07-17T08:34:50.1537841Z * [new branch] gh/XilunWu/157/head -> origin/gh/XilunWu/157/head 2025-07-17T08:34:50.1541224Z * [new branch] gh/XilunWu/157/orig -> origin/gh/XilunWu/157/orig 2025-07-17T08:34:50.1545822Z * [new branch] gh/XilunWu/158/base -> origin/gh/XilunWu/158/base 2025-07-17T08:34:50.1549134Z * [new branch] gh/XilunWu/158/head -> origin/gh/XilunWu/158/head 2025-07-17T08:34:50.1552516Z * [new branch] gh/XilunWu/158/orig -> origin/gh/XilunWu/158/orig 2025-07-17T08:34:50.1558596Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-07-17T08:34:50.1561825Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-07-17T08:34:50.1565291Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-07-17T08:34:50.1569930Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-07-17T08:34:50.1573426Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-07-17T08:34:50.1576945Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-07-17T08:34:50.1581369Z * [new branch] gh/XuehaiPan/189/base -> origin/gh/XuehaiPan/189/base 2025-07-17T08:34:50.1584687Z * [new branch] gh/XuehaiPan/189/head -> origin/gh/XuehaiPan/189/head 2025-07-17T08:34:50.1588019Z * [new branch] gh/XuehaiPan/189/orig -> origin/gh/XuehaiPan/189/orig 2025-07-17T08:34:50.1592571Z * [new branch] gh/XuehaiPan/226/base -> origin/gh/XuehaiPan/226/base 2025-07-17T08:34:50.1596181Z * [new branch] gh/XuehaiPan/226/head -> origin/gh/XuehaiPan/226/head 2025-07-17T08:34:50.1599418Z * [new branch] gh/XuehaiPan/226/orig -> origin/gh/XuehaiPan/226/orig 2025-07-17T08:34:50.1603973Z * [new branch] gh/XuehaiPan/227/base -> origin/gh/XuehaiPan/227/base 2025-07-17T08:34:50.1607316Z * [new branch] gh/XuehaiPan/227/head -> origin/gh/XuehaiPan/227/head 2025-07-17T08:34:50.1610655Z * [new branch] gh/XuehaiPan/227/orig -> origin/gh/XuehaiPan/227/orig 2025-07-17T08:34:50.1615173Z * [new branch] gh/XuehaiPan/228/base -> origin/gh/XuehaiPan/228/base 2025-07-17T08:34:50.1618603Z * [new branch] gh/XuehaiPan/228/head -> origin/gh/XuehaiPan/228/head 2025-07-17T08:34:50.1621941Z * [new branch] gh/XuehaiPan/228/orig -> origin/gh/XuehaiPan/228/orig 2025-07-17T08:34:50.1626627Z * [new branch] gh/XuehaiPan/229/base -> origin/gh/XuehaiPan/229/base 2025-07-17T08:34:50.1630011Z * [new branch] gh/XuehaiPan/229/head -> origin/gh/XuehaiPan/229/head 2025-07-17T08:34:50.1633348Z * [new branch] gh/XuehaiPan/229/orig -> origin/gh/XuehaiPan/229/orig 2025-07-17T08:34:50.1638302Z * [new branch] gh/XuehaiPan/230/base -> origin/gh/XuehaiPan/230/base 2025-07-17T08:34:50.1641750Z * [new branch] gh/XuehaiPan/230/head -> origin/gh/XuehaiPan/230/head 2025-07-17T08:34:50.1644960Z * [new branch] gh/XuehaiPan/230/orig -> origin/gh/XuehaiPan/230/orig 2025-07-17T08:34:50.1662782Z * [new branch] gh/XuehaiPan/231/base -> origin/gh/XuehaiPan/231/base 2025-07-17T08:34:50.1663187Z * [new branch] gh/XuehaiPan/231/head -> origin/gh/XuehaiPan/231/head 2025-07-17T08:34:50.1663574Z * [new branch] gh/XuehaiPan/231/orig -> origin/gh/XuehaiPan/231/orig 2025-07-17T08:34:50.1664425Z * [new branch] gh/XuehaiPan/232/base -> origin/gh/XuehaiPan/232/base 2025-07-17T08:34:50.1665149Z * [new branch] gh/XuehaiPan/232/head -> origin/gh/XuehaiPan/232/head 2025-07-17T08:34:50.1667615Z * [new branch] gh/XuehaiPan/232/orig -> origin/gh/XuehaiPan/232/orig 2025-07-17T08:34:50.1672553Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-07-17T08:34:50.1676024Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-07-17T08:34:50.1679380Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-07-17T08:34:50.1683925Z * [new branch] gh/XuehaiPan/251/base -> origin/gh/XuehaiPan/251/base 2025-07-17T08:34:50.1687575Z * [new branch] gh/XuehaiPan/251/head -> origin/gh/XuehaiPan/251/head 2025-07-17T08:34:50.1690884Z * [new branch] gh/XuehaiPan/251/orig -> origin/gh/XuehaiPan/251/orig 2025-07-17T08:34:50.1695345Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-07-17T08:34:50.1698687Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-07-17T08:34:50.1702119Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-07-17T08:34:50.1706639Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-07-17T08:34:50.1710059Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-07-17T08:34:50.1713405Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-07-17T08:34:50.1718144Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-07-17T08:34:50.1721451Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-07-17T08:34:50.1724761Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-07-17T08:34:50.1729498Z * [new branch] gh/XuehaiPan/257/base -> origin/gh/XuehaiPan/257/base 2025-07-17T08:34:50.1732870Z * [new branch] gh/XuehaiPan/257/head -> origin/gh/XuehaiPan/257/head 2025-07-17T08:34:50.1736176Z * [new branch] gh/XuehaiPan/257/orig -> origin/gh/XuehaiPan/257/orig 2025-07-17T08:34:50.1740855Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-07-17T08:34:50.1744037Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-07-17T08:34:50.1747344Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-07-17T08:34:50.1751929Z * [new branch] gh/XuehaiPan/279/base -> origin/gh/XuehaiPan/279/base 2025-07-17T08:34:50.1755803Z * [new branch] gh/XuehaiPan/279/head -> origin/gh/XuehaiPan/279/head 2025-07-17T08:34:50.1759143Z * [new branch] gh/XuehaiPan/279/orig -> origin/gh/XuehaiPan/279/orig 2025-07-17T08:34:50.1763660Z * [new branch] gh/XuehaiPan/281/base -> origin/gh/XuehaiPan/281/base 2025-07-17T08:34:50.1767041Z * [new branch] gh/XuehaiPan/281/head -> origin/gh/XuehaiPan/281/head 2025-07-17T08:34:50.1770529Z * [new branch] gh/XuehaiPan/281/orig -> origin/gh/XuehaiPan/281/orig 2025-07-17T08:34:50.1774931Z * [new branch] gh/XuehaiPan/282/base -> origin/gh/XuehaiPan/282/base 2025-07-17T08:34:50.1778330Z * [new branch] gh/XuehaiPan/282/head -> origin/gh/XuehaiPan/282/head 2025-07-17T08:34:50.1781689Z * [new branch] gh/XuehaiPan/282/orig -> origin/gh/XuehaiPan/282/orig 2025-07-17T08:34:50.1786214Z * [new branch] gh/XuehaiPan/283/base -> origin/gh/XuehaiPan/283/base 2025-07-17T08:34:50.1789534Z * [new branch] gh/XuehaiPan/283/head -> origin/gh/XuehaiPan/283/head 2025-07-17T08:34:50.1792944Z * [new branch] gh/XuehaiPan/283/orig -> origin/gh/XuehaiPan/283/orig 2025-07-17T08:34:50.1797756Z * [new branch] gh/XuehaiPan/284/base -> origin/gh/XuehaiPan/284/base 2025-07-17T08:34:50.1801146Z * [new branch] gh/XuehaiPan/284/head -> origin/gh/XuehaiPan/284/head 2025-07-17T08:34:50.1804494Z * [new branch] gh/XuehaiPan/284/orig -> origin/gh/XuehaiPan/284/orig 2025-07-17T08:34:50.1809048Z * [new branch] gh/XuehaiPan/285/base -> origin/gh/XuehaiPan/285/base 2025-07-17T08:34:50.1812381Z * [new branch] gh/XuehaiPan/285/head -> origin/gh/XuehaiPan/285/head 2025-07-17T08:34:50.1815667Z * [new branch] gh/XuehaiPan/285/orig -> origin/gh/XuehaiPan/285/orig 2025-07-17T08:34:50.1820281Z * [new branch] gh/XuehaiPan/286/base -> origin/gh/XuehaiPan/286/base 2025-07-17T08:34:50.1824225Z * [new branch] gh/XuehaiPan/286/head -> origin/gh/XuehaiPan/286/head 2025-07-17T08:34:50.1827602Z * [new branch] gh/XuehaiPan/286/orig -> origin/gh/XuehaiPan/286/orig 2025-07-17T08:34:50.1832087Z * [new branch] gh/XuehaiPan/287/base -> origin/gh/XuehaiPan/287/base 2025-07-17T08:34:50.1835710Z * [new branch] gh/XuehaiPan/287/head -> origin/gh/XuehaiPan/287/head 2025-07-17T08:34:50.1838994Z * [new branch] gh/XuehaiPan/287/orig -> origin/gh/XuehaiPan/287/orig 2025-07-17T08:34:50.1843538Z * [new branch] gh/XuehaiPan/288/base -> origin/gh/XuehaiPan/288/base 2025-07-17T08:34:50.1847031Z * [new branch] gh/XuehaiPan/288/head -> origin/gh/XuehaiPan/288/head 2025-07-17T08:34:50.1850336Z * [new branch] gh/XuehaiPan/288/orig -> origin/gh/XuehaiPan/288/orig 2025-07-17T08:34:50.1854809Z * [new branch] gh/XuehaiPan/289/base -> origin/gh/XuehaiPan/289/base 2025-07-17T08:34:50.1858130Z * [new branch] gh/XuehaiPan/289/head -> origin/gh/XuehaiPan/289/head 2025-07-17T08:34:50.1861663Z * [new branch] gh/XuehaiPan/289/orig -> origin/gh/XuehaiPan/289/orig 2025-07-17T08:34:50.1866261Z * [new branch] gh/XuehaiPan/290/base -> origin/gh/XuehaiPan/290/base 2025-07-17T08:34:50.1869595Z * [new branch] gh/XuehaiPan/290/head -> origin/gh/XuehaiPan/290/head 2025-07-17T08:34:50.1873027Z * [new branch] gh/XuehaiPan/290/orig -> origin/gh/XuehaiPan/290/orig 2025-07-17T08:34:50.1878085Z * [new branch] gh/XuehaiPan/291/base -> origin/gh/XuehaiPan/291/base 2025-07-17T08:34:50.1881779Z * [new branch] gh/XuehaiPan/291/head -> origin/gh/XuehaiPan/291/head 2025-07-17T08:34:50.1884655Z * [new branch] gh/XuehaiPan/291/orig -> origin/gh/XuehaiPan/291/orig 2025-07-17T08:34:50.1889138Z * [new branch] gh/XuehaiPan/292/base -> origin/gh/XuehaiPan/292/base 2025-07-17T08:34:50.1892570Z * [new branch] gh/XuehaiPan/292/head -> origin/gh/XuehaiPan/292/head 2025-07-17T08:34:50.1895922Z * [new branch] gh/XuehaiPan/292/orig -> origin/gh/XuehaiPan/292/orig 2025-07-17T08:34:50.1900419Z * [new branch] gh/XuehaiPan/293/base -> origin/gh/XuehaiPan/293/base 2025-07-17T08:34:50.1903905Z * [new branch] gh/XuehaiPan/293/head -> origin/gh/XuehaiPan/293/head 2025-07-17T08:34:50.1907078Z * [new branch] gh/XuehaiPan/293/orig -> origin/gh/XuehaiPan/293/orig 2025-07-17T08:34:50.1911751Z * [new branch] gh/XuehaiPan/294/base -> origin/gh/XuehaiPan/294/base 2025-07-17T08:34:50.1915190Z * [new branch] gh/XuehaiPan/294/head -> origin/gh/XuehaiPan/294/head 2025-07-17T08:34:50.1918600Z * [new branch] gh/XuehaiPan/294/orig -> origin/gh/XuehaiPan/294/orig 2025-07-17T08:34:50.1923126Z * [new branch] gh/XuehaiPan/295/base -> origin/gh/XuehaiPan/295/base 2025-07-17T08:34:50.1926457Z * [new branch] gh/XuehaiPan/295/head -> origin/gh/XuehaiPan/295/head 2025-07-17T08:34:50.1930072Z * [new branch] gh/XuehaiPan/295/orig -> origin/gh/XuehaiPan/295/orig 2025-07-17T08:34:50.1934336Z * [new branch] gh/XuehaiPan/296/base -> origin/gh/XuehaiPan/296/base 2025-07-17T08:34:50.1937778Z * [new branch] gh/XuehaiPan/296/head -> origin/gh/XuehaiPan/296/head 2025-07-17T08:34:50.1941096Z * [new branch] gh/XuehaiPan/296/orig -> origin/gh/XuehaiPan/296/orig 2025-07-17T08:34:50.1945727Z * [new branch] gh/XuehaiPan/297/base -> origin/gh/XuehaiPan/297/base 2025-07-17T08:34:50.1949052Z * [new branch] gh/XuehaiPan/297/head -> origin/gh/XuehaiPan/297/head 2025-07-17T08:34:50.1952376Z * [new branch] gh/XuehaiPan/297/orig -> origin/gh/XuehaiPan/297/orig 2025-07-17T08:34:50.1957223Z * [new branch] gh/XuehaiPan/298/base -> origin/gh/XuehaiPan/298/base 2025-07-17T08:34:50.1960536Z * [new branch] gh/XuehaiPan/298/head -> origin/gh/XuehaiPan/298/head 2025-07-17T08:34:50.1964131Z * [new branch] gh/XuehaiPan/298/orig -> origin/gh/XuehaiPan/298/orig 2025-07-17T08:34:50.1968684Z * [new branch] gh/XuehaiPan/299/base -> origin/gh/XuehaiPan/299/base 2025-07-17T08:34:50.1971986Z * [new branch] gh/XuehaiPan/299/head -> origin/gh/XuehaiPan/299/head 2025-07-17T08:34:50.1975308Z * [new branch] gh/XuehaiPan/299/orig -> origin/gh/XuehaiPan/299/orig 2025-07-17T08:34:50.1979808Z * [new branch] gh/XuehaiPan/300/base -> origin/gh/XuehaiPan/300/base 2025-07-17T08:34:50.1983264Z * [new branch] gh/XuehaiPan/300/head -> origin/gh/XuehaiPan/300/head 2025-07-17T08:34:50.1986665Z * [new branch] gh/XuehaiPan/300/orig -> origin/gh/XuehaiPan/300/orig 2025-07-17T08:34:50.1991173Z * [new branch] gh/XuehaiPan/301/base -> origin/gh/XuehaiPan/301/base 2025-07-17T08:34:50.1994653Z * [new branch] gh/XuehaiPan/301/head -> origin/gh/XuehaiPan/301/head 2025-07-17T08:34:50.1998122Z * [new branch] gh/XuehaiPan/301/orig -> origin/gh/XuehaiPan/301/orig 2025-07-17T08:34:50.2002762Z * [new branch] gh/XuehaiPan/302/base -> origin/gh/XuehaiPan/302/base 2025-07-17T08:34:50.2006108Z * [new branch] gh/XuehaiPan/302/head -> origin/gh/XuehaiPan/302/head 2025-07-17T08:34:50.2009432Z * [new branch] gh/XuehaiPan/302/orig -> origin/gh/XuehaiPan/302/orig 2025-07-17T08:34:50.2013934Z * [new branch] gh/XuehaiPan/303/base -> origin/gh/XuehaiPan/303/base 2025-07-17T08:34:50.2017285Z * [new branch] gh/XuehaiPan/303/head -> origin/gh/XuehaiPan/303/head 2025-07-17T08:34:50.2020616Z * [new branch] gh/XuehaiPan/303/orig -> origin/gh/XuehaiPan/303/orig 2025-07-17T08:34:50.2025182Z * [new branch] gh/XuehaiPan/304/base -> origin/gh/XuehaiPan/304/base 2025-07-17T08:34:50.2029064Z * [new branch] gh/XuehaiPan/304/head -> origin/gh/XuehaiPan/304/head 2025-07-17T08:34:50.2032020Z * [new branch] gh/XuehaiPan/304/orig -> origin/gh/XuehaiPan/304/orig 2025-07-17T08:34:50.2036873Z * [new branch] gh/XuehaiPan/305/base -> origin/gh/XuehaiPan/305/base 2025-07-17T08:34:50.2039998Z * [new branch] gh/XuehaiPan/305/head -> origin/gh/XuehaiPan/305/head 2025-07-17T08:34:50.2043353Z * [new branch] gh/XuehaiPan/305/orig -> origin/gh/XuehaiPan/305/orig 2025-07-17T08:34:50.2047894Z * [new branch] gh/XuehaiPan/306/base -> origin/gh/XuehaiPan/306/base 2025-07-17T08:34:50.2051467Z * [new branch] gh/XuehaiPan/306/head -> origin/gh/XuehaiPan/306/head 2025-07-17T08:34:50.2054887Z * [new branch] gh/XuehaiPan/306/orig -> origin/gh/XuehaiPan/306/orig 2025-07-17T08:34:50.2059264Z * [new branch] gh/XuehaiPan/307/base -> origin/gh/XuehaiPan/307/base 2025-07-17T08:34:50.2062585Z * [new branch] gh/XuehaiPan/307/head -> origin/gh/XuehaiPan/307/head 2025-07-17T08:34:50.2065902Z * [new branch] gh/XuehaiPan/307/orig -> origin/gh/XuehaiPan/307/orig 2025-07-17T08:34:50.2070555Z * [new branch] gh/XuehaiPan/308/base -> origin/gh/XuehaiPan/308/base 2025-07-17T08:34:50.2074062Z * [new branch] gh/XuehaiPan/308/head -> origin/gh/XuehaiPan/308/head 2025-07-17T08:34:50.2077557Z * [new branch] gh/XuehaiPan/308/orig -> origin/gh/XuehaiPan/308/orig 2025-07-17T08:34:50.2082019Z * [new branch] gh/XuehaiPan/309/base -> origin/gh/XuehaiPan/309/base 2025-07-17T08:34:50.2085380Z * [new branch] gh/XuehaiPan/309/head -> origin/gh/XuehaiPan/309/head 2025-07-17T08:34:50.2088721Z * [new branch] gh/XuehaiPan/309/orig -> origin/gh/XuehaiPan/309/orig 2025-07-17T08:34:50.2093375Z * [new branch] gh/XuehaiPan/310/base -> origin/gh/XuehaiPan/310/base 2025-07-17T08:34:50.2096724Z * [new branch] gh/XuehaiPan/310/head -> origin/gh/XuehaiPan/310/head 2025-07-17T08:34:50.2100092Z * [new branch] gh/XuehaiPan/310/orig -> origin/gh/XuehaiPan/310/orig 2025-07-17T08:34:50.2104691Z * [new branch] gh/XuehaiPan/311/base -> origin/gh/XuehaiPan/311/base 2025-07-17T08:34:50.2108041Z * [new branch] gh/XuehaiPan/311/head -> origin/gh/XuehaiPan/311/head 2025-07-17T08:34:50.2111384Z * [new branch] gh/XuehaiPan/311/orig -> origin/gh/XuehaiPan/311/orig 2025-07-17T08:34:50.2116255Z * [new branch] gh/XuehaiPan/312/base -> origin/gh/XuehaiPan/312/base 2025-07-17T08:34:50.2119624Z * [new branch] gh/XuehaiPan/312/head -> origin/gh/XuehaiPan/312/head 2025-07-17T08:34:50.2122994Z * [new branch] gh/XuehaiPan/312/orig -> origin/gh/XuehaiPan/312/orig 2025-07-17T08:34:50.2127563Z * [new branch] gh/XuehaiPan/313/base -> origin/gh/XuehaiPan/313/base 2025-07-17T08:34:50.2130939Z * [new branch] gh/XuehaiPan/313/head -> origin/gh/XuehaiPan/313/head 2025-07-17T08:34:50.2134287Z * [new branch] gh/XuehaiPan/313/orig -> origin/gh/XuehaiPan/313/orig 2025-07-17T08:34:50.2138881Z * [new branch] gh/XuehaiPan/314/base -> origin/gh/XuehaiPan/314/base 2025-07-17T08:34:50.2142267Z * [new branch] gh/XuehaiPan/314/head -> origin/gh/XuehaiPan/314/head 2025-07-17T08:34:50.2145586Z * [new branch] gh/XuehaiPan/314/orig -> origin/gh/XuehaiPan/314/orig 2025-07-17T08:34:50.2150133Z * [new branch] gh/XuehaiPan/315/base -> origin/gh/XuehaiPan/315/base 2025-07-17T08:34:50.2153469Z * [new branch] gh/XuehaiPan/315/head -> origin/gh/XuehaiPan/315/head 2025-07-17T08:34:50.2157068Z * [new branch] gh/XuehaiPan/315/orig -> origin/gh/XuehaiPan/315/orig 2025-07-17T08:34:50.2161528Z * [new branch] gh/XuehaiPan/316/base -> origin/gh/XuehaiPan/316/base 2025-07-17T08:34:50.2164946Z * [new branch] gh/XuehaiPan/316/head -> origin/gh/XuehaiPan/316/head 2025-07-17T08:34:50.2168457Z * [new branch] gh/XuehaiPan/316/orig -> origin/gh/XuehaiPan/316/orig 2025-07-17T08:34:50.2172918Z * [new branch] gh/XuehaiPan/317/base -> origin/gh/XuehaiPan/317/base 2025-07-17T08:34:50.2176191Z * [new branch] gh/XuehaiPan/317/head -> origin/gh/XuehaiPan/317/head 2025-07-17T08:34:50.2179542Z * [new branch] gh/XuehaiPan/317/orig -> origin/gh/XuehaiPan/317/orig 2025-07-17T08:34:50.2184266Z * [new branch] gh/XuehaiPan/318/base -> origin/gh/XuehaiPan/318/base 2025-07-17T08:34:50.2187577Z * [new branch] gh/XuehaiPan/318/head -> origin/gh/XuehaiPan/318/head 2025-07-17T08:34:50.2190926Z * [new branch] gh/XuehaiPan/318/orig -> origin/gh/XuehaiPan/318/orig 2025-07-17T08:34:50.2195739Z * [new branch] gh/XuehaiPan/319/base -> origin/gh/XuehaiPan/319/base 2025-07-17T08:34:50.2199033Z * [new branch] gh/XuehaiPan/319/head -> origin/gh/XuehaiPan/319/head 2025-07-17T08:34:50.2202457Z * [new branch] gh/XuehaiPan/319/orig -> origin/gh/XuehaiPan/319/orig 2025-07-17T08:34:50.2207032Z * [new branch] gh/XuehaiPan/320/base -> origin/gh/XuehaiPan/320/base 2025-07-17T08:34:50.2210473Z * [new branch] gh/XuehaiPan/320/head -> origin/gh/XuehaiPan/320/head 2025-07-17T08:34:50.2213809Z * [new branch] gh/XuehaiPan/320/orig -> origin/gh/XuehaiPan/320/orig 2025-07-17T08:34:50.2218284Z * [new branch] gh/XuehaiPan/321/base -> origin/gh/XuehaiPan/321/base 2025-07-17T08:34:50.2221590Z * [new branch] gh/XuehaiPan/321/head -> origin/gh/XuehaiPan/321/head 2025-07-17T08:34:50.2224927Z * [new branch] gh/XuehaiPan/321/orig -> origin/gh/XuehaiPan/321/orig 2025-07-17T08:34:50.2229380Z * [new branch] gh/XuehaiPan/322/base -> origin/gh/XuehaiPan/322/base 2025-07-17T08:34:50.2232784Z * [new branch] gh/XuehaiPan/322/head -> origin/gh/XuehaiPan/322/head 2025-07-17T08:34:50.2236339Z * [new branch] gh/XuehaiPan/322/orig -> origin/gh/XuehaiPan/322/orig 2025-07-17T08:34:50.2240923Z * [new branch] gh/XuehaiPan/323/base -> origin/gh/XuehaiPan/323/base 2025-07-17T08:34:50.2244593Z * [new branch] gh/XuehaiPan/323/head -> origin/gh/XuehaiPan/323/head 2025-07-17T08:34:50.2247989Z * [new branch] gh/XuehaiPan/323/orig -> origin/gh/XuehaiPan/323/orig 2025-07-17T08:34:50.2252520Z * [new branch] gh/XuehaiPan/324/base -> origin/gh/XuehaiPan/324/base 2025-07-17T08:34:50.2255941Z * [new branch] gh/XuehaiPan/324/head -> origin/gh/XuehaiPan/324/head 2025-07-17T08:34:50.2259287Z * [new branch] gh/XuehaiPan/324/orig -> origin/gh/XuehaiPan/324/orig 2025-07-17T08:34:50.2263831Z * [new branch] gh/XuehaiPan/325/base -> origin/gh/XuehaiPan/325/base 2025-07-17T08:34:50.2267218Z * [new branch] gh/XuehaiPan/325/head -> origin/gh/XuehaiPan/325/head 2025-07-17T08:34:50.2270575Z * [new branch] gh/XuehaiPan/325/orig -> origin/gh/XuehaiPan/325/orig 2025-07-17T08:34:50.2275409Z * [new branch] gh/XuehaiPan/326/base -> origin/gh/XuehaiPan/326/base 2025-07-17T08:34:50.2278727Z * [new branch] gh/XuehaiPan/326/head -> origin/gh/XuehaiPan/326/head 2025-07-17T08:34:50.2282041Z * [new branch] gh/XuehaiPan/326/orig -> origin/gh/XuehaiPan/326/orig 2025-07-17T08:34:50.2286625Z * [new branch] gh/XuehaiPan/327/base -> origin/gh/XuehaiPan/327/base 2025-07-17T08:34:50.2289965Z * [new branch] gh/XuehaiPan/327/head -> origin/gh/XuehaiPan/327/head 2025-07-17T08:34:50.2293308Z * [new branch] gh/XuehaiPan/327/orig -> origin/gh/XuehaiPan/327/orig 2025-07-17T08:34:50.2298114Z * [new branch] gh/XuehaiPan/328/base -> origin/gh/XuehaiPan/328/base 2025-07-17T08:34:50.2301400Z * [new branch] gh/XuehaiPan/328/head -> origin/gh/XuehaiPan/328/head 2025-07-17T08:34:50.2304738Z * [new branch] gh/XuehaiPan/328/orig -> origin/gh/XuehaiPan/328/orig 2025-07-17T08:34:50.2309295Z * [new branch] gh/XuehaiPan/329/base -> origin/gh/XuehaiPan/329/base 2025-07-17T08:34:50.2312605Z * [new branch] gh/XuehaiPan/329/head -> origin/gh/XuehaiPan/329/head 2025-07-17T08:34:50.2316241Z * [new branch] gh/XuehaiPan/329/orig -> origin/gh/XuehaiPan/329/orig 2025-07-17T08:34:50.2320792Z * [new branch] gh/XuehaiPan/330/base -> origin/gh/XuehaiPan/330/base 2025-07-17T08:34:50.2324183Z * [new branch] gh/XuehaiPan/330/head -> origin/gh/XuehaiPan/330/head 2025-07-17T08:34:50.2327550Z * [new branch] gh/XuehaiPan/330/orig -> origin/gh/XuehaiPan/330/orig 2025-07-17T08:34:50.2332050Z * [new branch] gh/XuehaiPan/331/base -> origin/gh/XuehaiPan/331/base 2025-07-17T08:34:50.2335341Z * [new branch] gh/XuehaiPan/331/head -> origin/gh/XuehaiPan/331/head 2025-07-17T08:34:50.2338701Z * [new branch] gh/XuehaiPan/331/orig -> origin/gh/XuehaiPan/331/orig 2025-07-17T08:34:50.2343245Z * [new branch] gh/XuehaiPan/332/base -> origin/gh/XuehaiPan/332/base 2025-07-17T08:34:50.2346727Z * [new branch] gh/XuehaiPan/332/head -> origin/gh/XuehaiPan/332/head 2025-07-17T08:34:50.2350114Z * [new branch] gh/XuehaiPan/332/orig -> origin/gh/XuehaiPan/332/orig 2025-07-17T08:34:50.2355549Z * [new branch] gh/XuehaiPan/333/base -> origin/gh/XuehaiPan/333/base 2025-07-17T08:34:50.2359209Z * [new branch] gh/XuehaiPan/333/head -> origin/gh/XuehaiPan/333/head 2025-07-17T08:34:50.2362470Z * [new branch] gh/XuehaiPan/333/orig -> origin/gh/XuehaiPan/333/orig 2025-07-17T08:34:50.2367094Z * [new branch] gh/XuehaiPan/334/base -> origin/gh/XuehaiPan/334/base 2025-07-17T08:34:50.2370397Z * [new branch] gh/XuehaiPan/334/head -> origin/gh/XuehaiPan/334/head 2025-07-17T08:34:50.2373677Z * [new branch] gh/XuehaiPan/334/orig -> origin/gh/XuehaiPan/334/orig 2025-07-17T08:34:50.2378506Z * [new branch] gh/XuehaiPan/335/base -> origin/gh/XuehaiPan/335/base 2025-07-17T08:34:50.2381539Z * [new branch] gh/XuehaiPan/335/head -> origin/gh/XuehaiPan/335/head 2025-07-17T08:34:50.2384834Z * [new branch] gh/XuehaiPan/335/orig -> origin/gh/XuehaiPan/335/orig 2025-07-17T08:34:50.2389577Z * [new branch] gh/XuehaiPan/336/base -> origin/gh/XuehaiPan/336/base 2025-07-17T08:34:50.2392771Z * [new branch] gh/XuehaiPan/336/head -> origin/gh/XuehaiPan/336/head 2025-07-17T08:34:50.2397530Z * [new branch] gh/XuehaiPan/336/orig -> origin/gh/XuehaiPan/336/orig 2025-07-17T08:34:50.2402676Z * [new branch] gh/XuehaiPan/337/base -> origin/gh/XuehaiPan/337/base 2025-07-17T08:34:50.2406088Z * [new branch] gh/XuehaiPan/337/head -> origin/gh/XuehaiPan/337/head 2025-07-17T08:34:50.2409490Z * [new branch] gh/XuehaiPan/337/orig -> origin/gh/XuehaiPan/337/orig 2025-07-17T08:34:50.2414036Z * [new branch] gh/XuehaiPan/338/base -> origin/gh/XuehaiPan/338/base 2025-07-17T08:34:50.2417357Z * [new branch] gh/XuehaiPan/338/head -> origin/gh/XuehaiPan/338/head 2025-07-17T08:34:50.2420699Z * [new branch] gh/XuehaiPan/338/orig -> origin/gh/XuehaiPan/338/orig 2025-07-17T08:34:50.2425341Z * [new branch] gh/XuehaiPan/339/base -> origin/gh/XuehaiPan/339/base 2025-07-17T08:34:50.2428723Z * [new branch] gh/XuehaiPan/339/head -> origin/gh/XuehaiPan/339/head 2025-07-17T08:34:50.2432390Z * [new branch] gh/XuehaiPan/339/orig -> origin/gh/XuehaiPan/339/orig 2025-07-17T08:34:50.2436995Z * [new branch] gh/XuehaiPan/340/base -> origin/gh/XuehaiPan/340/base 2025-07-17T08:34:50.2440313Z * [new branch] gh/XuehaiPan/340/head -> origin/gh/XuehaiPan/340/head 2025-07-17T08:34:50.2443686Z * [new branch] gh/XuehaiPan/340/orig -> origin/gh/XuehaiPan/340/orig 2025-07-17T08:34:50.2448221Z * [new branch] gh/XuehaiPan/341/base -> origin/gh/XuehaiPan/341/base 2025-07-17T08:34:50.2451587Z * [new branch] gh/XuehaiPan/341/head -> origin/gh/XuehaiPan/341/head 2025-07-17T08:34:50.2454905Z * [new branch] gh/XuehaiPan/341/orig -> origin/gh/XuehaiPan/341/orig 2025-07-17T08:34:50.2459563Z * [new branch] gh/XuehaiPan/342/base -> origin/gh/XuehaiPan/342/base 2025-07-17T08:34:50.2462897Z * [new branch] gh/XuehaiPan/342/head -> origin/gh/XuehaiPan/342/head 2025-07-17T08:34:50.2466264Z * [new branch] gh/XuehaiPan/342/orig -> origin/gh/XuehaiPan/342/orig 2025-07-17T08:34:50.2470932Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-07-17T08:34:50.2474434Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-07-17T08:34:50.2477909Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-07-17T08:34:50.2482413Z * [new branch] gh/XuehaiPan/344/base -> origin/gh/XuehaiPan/344/base 2025-07-17T08:34:50.2485748Z * [new branch] gh/XuehaiPan/344/head -> origin/gh/XuehaiPan/344/head 2025-07-17T08:34:50.2489047Z * [new branch] gh/XuehaiPan/344/orig -> origin/gh/XuehaiPan/344/orig 2025-07-17T08:34:50.2493560Z * [new branch] gh/XuehaiPan/345/base -> origin/gh/XuehaiPan/345/base 2025-07-17T08:34:50.2496892Z * [new branch] gh/XuehaiPan/345/head -> origin/gh/XuehaiPan/345/head 2025-07-17T08:34:50.2500199Z * [new branch] gh/XuehaiPan/345/orig -> origin/gh/XuehaiPan/345/orig 2025-07-17T08:34:50.2504741Z * [new branch] gh/XuehaiPan/346/base -> origin/gh/XuehaiPan/346/base 2025-07-17T08:34:50.2508041Z * [new branch] gh/XuehaiPan/346/head -> origin/gh/XuehaiPan/346/head 2025-07-17T08:34:50.2511418Z * [new branch] gh/XuehaiPan/346/orig -> origin/gh/XuehaiPan/346/orig 2025-07-17T08:34:50.2516278Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-07-17T08:34:50.2519545Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-07-17T08:34:50.2522886Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-07-17T08:34:50.2527726Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-07-17T08:34:50.2531042Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-07-17T08:34:50.2534431Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-07-17T08:34:50.2539102Z * [new branch] gh/XuehaiPan/349/base -> origin/gh/XuehaiPan/349/base 2025-07-17T08:34:50.2542455Z * [new branch] gh/XuehaiPan/349/head -> origin/gh/XuehaiPan/349/head 2025-07-17T08:34:50.2545840Z * [new branch] gh/XuehaiPan/349/orig -> origin/gh/XuehaiPan/349/orig 2025-07-17T08:34:50.2550478Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-07-17T08:34:50.2553827Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-07-17T08:34:50.2557456Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-07-17T08:34:50.2562042Z * [new branch] gh/XuehaiPan/351/base -> origin/gh/XuehaiPan/351/base 2025-07-17T08:34:50.2565547Z * [new branch] gh/XuehaiPan/351/head -> origin/gh/XuehaiPan/351/head 2025-07-17T08:34:50.2568731Z * [new branch] gh/XuehaiPan/351/orig -> origin/gh/XuehaiPan/351/orig 2025-07-17T08:34:50.2573302Z * [new branch] gh/XuehaiPan/352/base -> origin/gh/XuehaiPan/352/base 2025-07-17T08:34:50.2576684Z * [new branch] gh/XuehaiPan/352/head -> origin/gh/XuehaiPan/352/head 2025-07-17T08:34:50.2579929Z * [new branch] gh/XuehaiPan/352/orig -> origin/gh/XuehaiPan/352/orig 2025-07-17T08:34:50.2584530Z * [new branch] gh/XuehaiPan/353/base -> origin/gh/XuehaiPan/353/base 2025-07-17T08:34:50.2587849Z * [new branch] gh/XuehaiPan/353/head -> origin/gh/XuehaiPan/353/head 2025-07-17T08:34:50.2591191Z * [new branch] gh/XuehaiPan/353/orig -> origin/gh/XuehaiPan/353/orig 2025-07-17T08:34:50.2596249Z * [new branch] gh/XuehaiPan/354/base -> origin/gh/XuehaiPan/354/base 2025-07-17T08:34:50.2599843Z * [new branch] gh/XuehaiPan/354/head -> origin/gh/XuehaiPan/354/head 2025-07-17T08:34:50.2603141Z * [new branch] gh/XuehaiPan/354/orig -> origin/gh/XuehaiPan/354/orig 2025-07-17T08:34:50.2607932Z * [new branch] gh/XuehaiPan/355/base -> origin/gh/XuehaiPan/355/base 2025-07-17T08:34:50.2611190Z * [new branch] gh/XuehaiPan/355/head -> origin/gh/XuehaiPan/355/head 2025-07-17T08:34:50.2614584Z * [new branch] gh/XuehaiPan/355/orig -> origin/gh/XuehaiPan/355/orig 2025-07-17T08:34:50.2619182Z * [new branch] gh/XuehaiPan/356/base -> origin/gh/XuehaiPan/356/base 2025-07-17T08:34:50.2622692Z * [new branch] gh/XuehaiPan/356/head -> origin/gh/XuehaiPan/356/head 2025-07-17T08:34:50.2626066Z * [new branch] gh/XuehaiPan/356/orig -> origin/gh/XuehaiPan/356/orig 2025-07-17T08:34:50.2630663Z * [new branch] gh/XuehaiPan/357/base -> origin/gh/XuehaiPan/357/base 2025-07-17T08:34:50.2634058Z * [new branch] gh/XuehaiPan/357/head -> origin/gh/XuehaiPan/357/head 2025-07-17T08:34:50.2638122Z * [new branch] gh/XuehaiPan/357/orig -> origin/gh/XuehaiPan/357/orig 2025-07-17T08:34:50.2642239Z * [new branch] gh/XuehaiPan/358/base -> origin/gh/XuehaiPan/358/base 2025-07-17T08:34:50.2645550Z * [new branch] gh/XuehaiPan/358/head -> origin/gh/XuehaiPan/358/head 2025-07-17T08:34:50.2648883Z * [new branch] gh/XuehaiPan/358/orig -> origin/gh/XuehaiPan/358/orig 2025-07-17T08:34:50.2653492Z * [new branch] gh/XuehaiPan/359/base -> origin/gh/XuehaiPan/359/base 2025-07-17T08:34:50.2656822Z * [new branch] gh/XuehaiPan/359/head -> origin/gh/XuehaiPan/359/head 2025-07-17T08:34:50.2660132Z * [new branch] gh/XuehaiPan/359/orig -> origin/gh/XuehaiPan/359/orig 2025-07-17T08:34:50.2664629Z * [new branch] gh/XuehaiPan/360/base -> origin/gh/XuehaiPan/360/base 2025-07-17T08:34:50.2667998Z * [new branch] gh/XuehaiPan/360/head -> origin/gh/XuehaiPan/360/head 2025-07-17T08:34:50.2671325Z * [new branch] gh/XuehaiPan/360/orig -> origin/gh/XuehaiPan/360/orig 2025-07-17T08:34:50.2676051Z * [new branch] gh/XuehaiPan/361/base -> origin/gh/XuehaiPan/361/base 2025-07-17T08:34:50.2679452Z * [new branch] gh/XuehaiPan/361/head -> origin/gh/XuehaiPan/361/head 2025-07-17T08:34:50.2682763Z * [new branch] gh/XuehaiPan/361/orig -> origin/gh/XuehaiPan/361/orig 2025-07-17T08:34:50.2687283Z * [new branch] gh/XuehaiPan/362/base -> origin/gh/XuehaiPan/362/base 2025-07-17T08:34:50.2690598Z * [new branch] gh/XuehaiPan/362/head -> origin/gh/XuehaiPan/362/head 2025-07-17T08:34:50.2694155Z * [new branch] gh/XuehaiPan/362/orig -> origin/gh/XuehaiPan/362/orig 2025-07-17T08:34:50.2698636Z * [new branch] gh/XuehaiPan/363/base -> origin/gh/XuehaiPan/363/base 2025-07-17T08:34:50.2701971Z * [new branch] gh/XuehaiPan/363/head -> origin/gh/XuehaiPan/363/head 2025-07-17T08:34:50.2705328Z * [new branch] gh/XuehaiPan/363/orig -> origin/gh/XuehaiPan/363/orig 2025-07-17T08:34:50.2709909Z * [new branch] gh/XuehaiPan/364/base -> origin/gh/XuehaiPan/364/base 2025-07-17T08:34:50.2713255Z * [new branch] gh/XuehaiPan/364/head -> origin/gh/XuehaiPan/364/head 2025-07-17T08:34:50.2716808Z * [new branch] gh/XuehaiPan/364/orig -> origin/gh/XuehaiPan/364/orig 2025-07-17T08:34:50.2721363Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-07-17T08:34:50.2724655Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-07-17T08:34:50.2728000Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-07-17T08:34:50.2732692Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-07-17T08:34:50.2736014Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-07-17T08:34:50.2740528Z * [new branch] gh/XuehaiPan/367/base -> origin/gh/XuehaiPan/367/base 2025-07-17T08:34:50.2743971Z * [new branch] gh/XuehaiPan/367/head -> origin/gh/XuehaiPan/367/head 2025-07-17T08:34:50.2748628Z * [new branch] gh/XuehaiPan/368/base -> origin/gh/XuehaiPan/368/base 2025-07-17T08:34:50.2751940Z * [new branch] gh/XuehaiPan/368/head -> origin/gh/XuehaiPan/368/head 2025-07-17T08:34:50.2755458Z * [new branch] gh/XuehaiPan/368/orig -> origin/gh/XuehaiPan/368/orig 2025-07-17T08:34:50.2759966Z * [new branch] gh/XuehaiPan/369/base -> origin/gh/XuehaiPan/369/base 2025-07-17T08:34:50.2763304Z * [new branch] gh/XuehaiPan/369/head -> origin/gh/XuehaiPan/369/head 2025-07-17T08:34:50.2766606Z * [new branch] gh/XuehaiPan/369/orig -> origin/gh/XuehaiPan/369/orig 2025-07-17T08:34:50.2772270Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-07-17T08:34:50.2775549Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-07-17T08:34:50.2778913Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-07-17T08:34:50.2783751Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-07-17T08:34:50.2786965Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-07-17T08:34:50.2791323Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-07-17T08:34:50.2794742Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-07-17T08:34:50.2799493Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-07-17T08:34:50.2802793Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-07-17T08:34:50.2807300Z * [new branch] gh/ZhiweiYan-96/56/base -> origin/gh/ZhiweiYan-96/56/base 2025-07-17T08:34:50.2810942Z * [new branch] gh/ZhiweiYan-96/56/head -> origin/gh/ZhiweiYan-96/56/head 2025-07-17T08:34:50.2814524Z * [new branch] gh/ZhiweiYan-96/56/orig -> origin/gh/ZhiweiYan-96/56/orig 2025-07-17T08:34:50.2820256Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-07-17T08:34:50.2822461Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-07-17T08:34:50.2827023Z * [new branch] gh/ZhiweiYan-96/64/base -> origin/gh/ZhiweiYan-96/64/base 2025-07-17T08:34:50.2830662Z * [new branch] gh/ZhiweiYan-96/64/head -> origin/gh/ZhiweiYan-96/64/head 2025-07-17T08:34:50.2833847Z * [new branch] gh/ZhiweiYan-96/64/orig -> origin/gh/ZhiweiYan-96/64/orig 2025-07-17T08:34:50.2839601Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-07-17T08:34:50.2842925Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-07-17T08:34:50.2847400Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-07-17T08:34:50.2850772Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-07-17T08:34:50.2855247Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-07-17T08:34:50.2858579Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-07-17T08:34:50.2861956Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-07-17T08:34:50.2867378Z * [new branch] gh/albanD/3/base -> origin/gh/albanD/3/base 2025-07-17T08:34:50.2870839Z * [new branch] gh/albanD/3/head -> origin/gh/albanD/3/head 2025-07-17T08:34:50.2874271Z * [new branch] gh/albanD/3/orig -> origin/gh/albanD/3/orig 2025-07-17T08:34:50.2879138Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-07-17T08:34:50.2882269Z * [new branch] gh/alexbrauckmann/paddedtensor_init -> origin/gh/alexbrauckmann/paddedtensor_init 2025-07-17T08:34:50.2885589Z * [new branch] gh/alexbrauckmann/paddedtensor_meta_init -> origin/gh/alexbrauckmann/paddedtensor_meta_init 2025-07-17T08:34:50.2890890Z * [new branch] gh/alexsamardzic/4/base -> origin/gh/alexsamardzic/4/base 2025-07-17T08:34:50.2894195Z * [new branch] gh/alexsamardzic/4/head -> origin/gh/alexsamardzic/4/head 2025-07-17T08:34:50.2897567Z * [new branch] gh/alexsamardzic/4/orig -> origin/gh/alexsamardzic/4/orig 2025-07-17T08:34:50.2902084Z * [new branch] gh/alexsamardzic/5/base -> origin/gh/alexsamardzic/5/base 2025-07-17T08:34:50.2905342Z * [new branch] gh/alexsamardzic/5/head -> origin/gh/alexsamardzic/5/head 2025-07-17T08:34:50.2908712Z * [new branch] gh/alexsamardzic/5/orig -> origin/gh/alexsamardzic/5/orig 2025-07-17T08:34:50.2913346Z * [new branch] gh/alexsamardzic/6/base -> origin/gh/alexsamardzic/6/base 2025-07-17T08:34:50.2917267Z * [new branch] gh/alexsamardzic/6/head -> origin/gh/alexsamardzic/6/head 2025-07-17T08:34:50.2920567Z * [new branch] gh/alexsamardzic/6/orig -> origin/gh/alexsamardzic/6/orig 2025-07-17T08:34:50.2925120Z * [new branch] gh/alexsamardzic/7/base -> origin/gh/alexsamardzic/7/base 2025-07-17T08:34:50.2928473Z * [new branch] gh/alexsamardzic/7/head -> origin/gh/alexsamardzic/7/head 2025-07-17T08:34:50.2931796Z * [new branch] gh/alexsamardzic/7/orig -> origin/gh/alexsamardzic/7/orig 2025-07-17T08:34:50.2937271Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-07-17T08:34:50.2940603Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-07-17T08:34:50.2943891Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-07-17T08:34:50.2949634Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-07-17T08:34:50.2953062Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-07-17T08:34:50.2956793Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-07-17T08:34:50.2961440Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-07-17T08:34:50.2965126Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-07-17T08:34:50.2968365Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-07-17T08:34:50.2973907Z * [new branch] gh/andyanwang/1/base -> origin/gh/andyanwang/1/base 2025-07-17T08:34:50.2977235Z * [new branch] gh/andyanwang/1/head -> origin/gh/andyanwang/1/head 2025-07-17T08:34:50.2980612Z * [new branch] gh/andyanwang/1/orig -> origin/gh/andyanwang/1/orig 2025-07-17T08:34:50.2985318Z * [new branch] gh/andyanwang/13/base -> origin/gh/andyanwang/13/base 2025-07-17T08:34:50.2988667Z * [new branch] gh/andyanwang/13/head -> origin/gh/andyanwang/13/head 2025-07-17T08:34:50.2991991Z * [new branch] gh/andyanwang/13/orig -> origin/gh/andyanwang/13/orig 2025-07-17T08:34:50.2996988Z * [new branch] gh/andyanwang/14/base -> origin/gh/andyanwang/14/base 2025-07-17T08:34:50.3000376Z * [new branch] gh/andyanwang/14/head -> origin/gh/andyanwang/14/head 2025-07-17T08:34:50.3003816Z * [new branch] gh/andyanwang/14/orig -> origin/gh/andyanwang/14/orig 2025-07-17T08:34:50.3008482Z * [new branch] gh/andyanwang/18/base -> origin/gh/andyanwang/18/base 2025-07-17T08:34:50.3012150Z * [new branch] gh/andyanwang/18/head -> origin/gh/andyanwang/18/head 2025-07-17T08:34:50.3015496Z * [new branch] gh/andyanwang/18/orig -> origin/gh/andyanwang/18/orig 2025-07-17T08:34:50.3020019Z * [new branch] gh/andyanwang/19/base -> origin/gh/andyanwang/19/base 2025-07-17T08:34:50.3023298Z * [new branch] gh/andyanwang/19/head -> origin/gh/andyanwang/19/head 2025-07-17T08:34:50.3026570Z * [new branch] gh/andyanwang/19/orig -> origin/gh/andyanwang/19/orig 2025-07-17T08:34:50.3031124Z * [new branch] gh/andyanwang/2/base -> origin/gh/andyanwang/2/base 2025-07-17T08:34:50.3034682Z * [new branch] gh/andyanwang/2/head -> origin/gh/andyanwang/2/head 2025-07-17T08:34:50.3038117Z * [new branch] gh/andyanwang/2/orig -> origin/gh/andyanwang/2/orig 2025-07-17T08:34:50.3042694Z * [new branch] gh/andyanwang/20/base -> origin/gh/andyanwang/20/base 2025-07-17T08:34:50.3046091Z * [new branch] gh/andyanwang/20/head -> origin/gh/andyanwang/20/head 2025-07-17T08:34:50.3049509Z * [new branch] gh/andyanwang/20/orig -> origin/gh/andyanwang/20/orig 2025-07-17T08:34:50.3054194Z * [new branch] gh/andyanwang/21/base -> origin/gh/andyanwang/21/base 2025-07-17T08:34:50.3057605Z * [new branch] gh/andyanwang/21/head -> origin/gh/andyanwang/21/head 2025-07-17T08:34:50.3061056Z * [new branch] gh/andyanwang/21/orig -> origin/gh/andyanwang/21/orig 2025-07-17T08:34:50.3065569Z * [new branch] gh/andyanwang/22/base -> origin/gh/andyanwang/22/base 2025-07-17T08:34:50.3068916Z * [new branch] gh/andyanwang/22/head -> origin/gh/andyanwang/22/head 2025-07-17T08:34:50.3072325Z * [new branch] gh/andyanwang/22/orig -> origin/gh/andyanwang/22/orig 2025-07-17T08:34:50.3077108Z * [new branch] gh/andyanwang/23/base -> origin/gh/andyanwang/23/base 2025-07-17T08:34:50.3080358Z * [new branch] gh/andyanwang/23/head -> origin/gh/andyanwang/23/head 2025-07-17T08:34:50.3083719Z * [new branch] gh/andyanwang/23/orig -> origin/gh/andyanwang/23/orig 2025-07-17T08:34:50.3088432Z * [new branch] gh/andyanwang/24/base -> origin/gh/andyanwang/24/base 2025-07-17T08:34:50.3091954Z * [new branch] gh/andyanwang/24/head -> origin/gh/andyanwang/24/head 2025-07-17T08:34:50.3095655Z * [new branch] gh/andyanwang/24/orig -> origin/gh/andyanwang/24/orig 2025-07-17T08:34:50.3100579Z * [new branch] gh/andyanwang/25/base -> origin/gh/andyanwang/25/base 2025-07-17T08:34:50.3103710Z * [new branch] gh/andyanwang/25/head -> origin/gh/andyanwang/25/head 2025-07-17T08:34:50.3107102Z * [new branch] gh/andyanwang/25/orig -> origin/gh/andyanwang/25/orig 2025-07-17T08:34:50.3111672Z * [new branch] gh/andyanwang/26/base -> origin/gh/andyanwang/26/base 2025-07-17T08:34:50.3115388Z * [new branch] gh/andyanwang/26/head -> origin/gh/andyanwang/26/head 2025-07-17T08:34:50.3119048Z * [new branch] gh/andyanwang/26/orig -> origin/gh/andyanwang/26/orig 2025-07-17T08:34:50.3123184Z * [new branch] gh/andyanwang/27/base -> origin/gh/andyanwang/27/base 2025-07-17T08:34:50.3126642Z * [new branch] gh/andyanwang/27/head -> origin/gh/andyanwang/27/head 2025-07-17T08:34:50.3129972Z * [new branch] gh/andyanwang/27/orig -> origin/gh/andyanwang/27/orig 2025-07-17T08:34:50.3134609Z * [new branch] gh/andyanwang/28/base -> origin/gh/andyanwang/28/base 2025-07-17T08:34:50.3138094Z * [new branch] gh/andyanwang/28/head -> origin/gh/andyanwang/28/head 2025-07-17T08:34:50.3141443Z * [new branch] gh/andyanwang/28/orig -> origin/gh/andyanwang/28/orig 2025-07-17T08:34:50.3146310Z * [new branch] gh/andyanwang/29/base -> origin/gh/andyanwang/29/base 2025-07-17T08:34:50.3149735Z * [new branch] gh/andyanwang/29/head -> origin/gh/andyanwang/29/head 2025-07-17T08:34:50.3153137Z * [new branch] gh/andyanwang/29/orig -> origin/gh/andyanwang/29/orig 2025-07-17T08:34:50.3160692Z * [new branch] gh/andyanwang/3/base -> origin/gh/andyanwang/3/base 2025-07-17T08:34:50.3164079Z * [new branch] gh/andyanwang/3/head -> origin/gh/andyanwang/3/head 2025-07-17T08:34:50.3167471Z * [new branch] gh/andyanwang/3/orig -> origin/gh/andyanwang/3/orig 2025-07-17T08:34:50.3172050Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-07-17T08:34:50.3175571Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-07-17T08:34:50.3180132Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-07-17T08:34:50.3183627Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-07-17T08:34:50.3188499Z * [new branch] gh/andyanwang/32/base -> origin/gh/andyanwang/32/base 2025-07-17T08:34:50.3192085Z * [new branch] gh/andyanwang/32/head -> origin/gh/andyanwang/32/head 2025-07-17T08:34:50.3195806Z * [new branch] gh/andyanwang/32/orig -> origin/gh/andyanwang/32/orig 2025-07-17T08:34:50.3200424Z * [new branch] gh/andyanwang/4/base -> origin/gh/andyanwang/4/base 2025-07-17T08:34:50.3203725Z * [new branch] gh/andyanwang/4/head -> origin/gh/andyanwang/4/head 2025-07-17T08:34:50.3207344Z * [new branch] gh/andyanwang/4/orig -> origin/gh/andyanwang/4/orig 2025-07-17T08:34:50.3212034Z * [new branch] gh/andyanwang/7/base -> origin/gh/andyanwang/7/base 2025-07-17T08:34:50.3215388Z * [new branch] gh/andyanwang/7/head -> origin/gh/andyanwang/7/head 2025-07-17T08:34:50.3218811Z * [new branch] gh/andyanwang/7/orig -> origin/gh/andyanwang/7/orig 2025-07-17T08:34:50.3224299Z * [new branch] gh/angelayi/100/base -> origin/gh/angelayi/100/base 2025-07-17T08:34:50.3227704Z * [new branch] gh/angelayi/100/head -> origin/gh/angelayi/100/head 2025-07-17T08:34:50.3231114Z * [new branch] gh/angelayi/100/orig -> origin/gh/angelayi/100/orig 2025-07-17T08:34:50.3236065Z * [new branch] gh/angelayi/101/base -> origin/gh/angelayi/101/base 2025-07-17T08:34:50.3239161Z * [new branch] gh/angelayi/101/head -> origin/gh/angelayi/101/head 2025-07-17T08:34:50.3242437Z * [new branch] gh/angelayi/101/orig -> origin/gh/angelayi/101/orig 2025-07-17T08:34:50.3247067Z * [new branch] gh/angelayi/102/base -> origin/gh/angelayi/102/base 2025-07-17T08:34:50.3250403Z * [new branch] gh/angelayi/102/head -> origin/gh/angelayi/102/head 2025-07-17T08:34:50.3253755Z * [new branch] gh/angelayi/102/orig -> origin/gh/angelayi/102/orig 2025-07-17T08:34:50.3258322Z * [new branch] gh/angelayi/103/base -> origin/gh/angelayi/103/base 2025-07-17T08:34:50.3261580Z * [new branch] gh/angelayi/103/head -> origin/gh/angelayi/103/head 2025-07-17T08:34:50.3264886Z * [new branch] gh/angelayi/103/orig -> origin/gh/angelayi/103/orig 2025-07-17T08:34:50.3269446Z * [new branch] gh/angelayi/104/base -> origin/gh/angelayi/104/base 2025-07-17T08:34:50.3272766Z * [new branch] gh/angelayi/104/head -> origin/gh/angelayi/104/head 2025-07-17T08:34:50.3276441Z * [new branch] gh/angelayi/104/orig -> origin/gh/angelayi/104/orig 2025-07-17T08:34:50.3280981Z * [new branch] gh/angelayi/82/base -> origin/gh/angelayi/82/base 2025-07-17T08:34:50.3284335Z * [new branch] gh/angelayi/82/head -> origin/gh/angelayi/82/head 2025-07-17T08:34:50.3287685Z * [new branch] gh/angelayi/82/orig -> origin/gh/angelayi/82/orig 2025-07-17T08:34:50.3292230Z * [new branch] gh/angelayi/97/base -> origin/gh/angelayi/97/base 2025-07-17T08:34:50.3295533Z * [new branch] gh/angelayi/97/head -> origin/gh/angelayi/97/head 2025-07-17T08:34:50.3298862Z * [new branch] gh/angelayi/97/orig -> origin/gh/angelayi/97/orig 2025-07-17T08:34:50.3304291Z * [new branch] gh/ani300/1/base -> origin/gh/ani300/1/base 2025-07-17T08:34:50.3307654Z * [new branch] gh/ani300/1/head -> origin/gh/ani300/1/head 2025-07-17T08:34:50.3310972Z * [new branch] gh/ani300/1/orig -> origin/gh/ani300/1/orig 2025-07-17T08:34:50.3316918Z * [new branch] gh/anijain2305/162/base -> origin/gh/anijain2305/162/base 2025-07-17T08:34:50.3320242Z * [new branch] gh/anijain2305/162/head -> origin/gh/anijain2305/162/head 2025-07-17T08:34:50.3324891Z * [new branch] gh/anijain2305/710/base -> origin/gh/anijain2305/710/base 2025-07-17T08:34:50.3328108Z * [new branch] gh/anijain2305/710/head -> origin/gh/anijain2305/710/head 2025-07-17T08:34:50.3331394Z * [new branch] gh/anijain2305/710/orig -> origin/gh/anijain2305/710/orig 2025-07-17T08:34:50.3336057Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-07-17T08:34:50.3339403Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-07-17T08:34:50.3342736Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-07-17T08:34:50.3347364Z * [new branch] gh/anijain2305/766/base -> origin/gh/anijain2305/766/base 2025-07-17T08:34:50.3350657Z * [new branch] gh/anijain2305/766/head -> origin/gh/anijain2305/766/head 2025-07-17T08:34:50.3354003Z * [new branch] gh/anijain2305/766/orig -> origin/gh/anijain2305/766/orig 2025-07-17T08:34:50.3358765Z * [new branch] gh/anijain2305/790/base -> origin/gh/anijain2305/790/base 2025-07-17T08:34:50.3362151Z * [new branch] gh/anijain2305/790/head -> origin/gh/anijain2305/790/head 2025-07-17T08:34:50.3365464Z * [new branch] gh/anijain2305/790/orig -> origin/gh/anijain2305/790/orig 2025-07-17T08:34:50.3370155Z * [new branch] gh/anijain2305/791/base -> origin/gh/anijain2305/791/base 2025-07-17T08:34:50.3373316Z * [new branch] gh/anijain2305/791/head -> origin/gh/anijain2305/791/head 2025-07-17T08:34:50.3376736Z * [new branch] gh/anijain2305/791/orig -> origin/gh/anijain2305/791/orig 2025-07-17T08:34:50.3381209Z * [new branch] gh/anijain2305/792/base -> origin/gh/anijain2305/792/base 2025-07-17T08:34:50.3384858Z * [new branch] gh/anijain2305/792/head -> origin/gh/anijain2305/792/head 2025-07-17T08:34:50.3388203Z * [new branch] gh/anijain2305/792/orig -> origin/gh/anijain2305/792/orig 2025-07-17T08:34:50.3392794Z * [new branch] gh/anijain2305/793/base -> origin/gh/anijain2305/793/base 2025-07-17T08:34:50.3397153Z * [new branch] gh/anijain2305/793/head -> origin/gh/anijain2305/793/head 2025-07-17T08:34:50.3400673Z * [new branch] gh/anijain2305/793/orig -> origin/gh/anijain2305/793/orig 2025-07-17T08:34:50.3405236Z * [new branch] gh/anijain2305/794/base -> origin/gh/anijain2305/794/base 2025-07-17T08:34:50.3408644Z * [new branch] gh/anijain2305/794/head -> origin/gh/anijain2305/794/head 2025-07-17T08:34:50.3411991Z * [new branch] gh/anijain2305/794/orig -> origin/gh/anijain2305/794/orig 2025-07-17T08:34:50.3416530Z * [new branch] gh/anijain2305/795/base -> origin/gh/anijain2305/795/base 2025-07-17T08:34:50.3419782Z * [new branch] gh/anijain2305/795/head -> origin/gh/anijain2305/795/head 2025-07-17T08:34:50.3423111Z * [new branch] gh/anijain2305/795/orig -> origin/gh/anijain2305/795/orig 2025-07-17T08:34:50.3427606Z * [new branch] gh/anijain2305/796/base -> origin/gh/anijain2305/796/base 2025-07-17T08:34:50.3430987Z * [new branch] gh/anijain2305/796/head -> origin/gh/anijain2305/796/head 2025-07-17T08:34:50.3434416Z * [new branch] gh/anijain2305/796/orig -> origin/gh/anijain2305/796/orig 2025-07-17T08:34:50.3438870Z * [new branch] gh/anijain2305/797/base -> origin/gh/anijain2305/797/base 2025-07-17T08:34:50.3442177Z * [new branch] gh/anijain2305/797/head -> origin/gh/anijain2305/797/head 2025-07-17T08:34:50.3445563Z * [new branch] gh/anijain2305/797/orig -> origin/gh/anijain2305/797/orig 2025-07-17T08:34:50.3450105Z * [new branch] gh/anijain2305/798/base -> origin/gh/anijain2305/798/base 2025-07-17T08:34:50.3453596Z * [new branch] gh/anijain2305/798/head -> origin/gh/anijain2305/798/head 2025-07-17T08:34:50.3458163Z * [new branch] gh/anijain2305/799/base -> origin/gh/anijain2305/799/base 2025-07-17T08:34:50.3461508Z * [new branch] gh/anijain2305/799/head -> origin/gh/anijain2305/799/head 2025-07-17T08:34:50.3466009Z * [new branch] gh/anijain2305/800/base -> origin/gh/anijain2305/800/base 2025-07-17T08:34:50.3469310Z * [new branch] gh/anijain2305/800/head -> origin/gh/anijain2305/800/head 2025-07-17T08:34:50.3473930Z * [new branch] gh/anijain2305/801/base -> origin/gh/anijain2305/801/base 2025-07-17T08:34:50.3477293Z * [new branch] gh/anijain2305/801/head -> origin/gh/anijain2305/801/head 2025-07-17T08:34:50.3480589Z * [new branch] gh/anijain2305/801/orig -> origin/gh/anijain2305/801/orig 2025-07-17T08:34:50.3485165Z * [new branch] gh/anijain2305/802/base -> origin/gh/anijain2305/802/base 2025-07-17T08:34:50.3488476Z * [new branch] gh/anijain2305/802/head -> origin/gh/anijain2305/802/head 2025-07-17T08:34:50.3491787Z * [new branch] gh/anijain2305/802/orig -> origin/gh/anijain2305/802/orig 2025-07-17T08:34:50.3496496Z * [new branch] gh/anijain2305/803/base -> origin/gh/anijain2305/803/base 2025-07-17T08:34:50.3499895Z * [new branch] gh/anijain2305/803/head -> origin/gh/anijain2305/803/head 2025-07-17T08:34:50.3503050Z * [new branch] gh/anijain2305/803/orig -> origin/gh/anijain2305/803/orig 2025-07-17T08:34:50.3507564Z * [new branch] gh/anijain2305/804/base -> origin/gh/anijain2305/804/base 2025-07-17T08:34:50.3510860Z * [new branch] gh/anijain2305/804/head -> origin/gh/anijain2305/804/head 2025-07-17T08:34:50.3514286Z * [new branch] gh/anijain2305/804/orig -> origin/gh/anijain2305/804/orig 2025-07-17T08:34:50.3518973Z * [new branch] gh/anijain2305/805/base -> origin/gh/anijain2305/805/base 2025-07-17T08:34:50.3522302Z * [new branch] gh/anijain2305/805/head -> origin/gh/anijain2305/805/head 2025-07-17T08:34:50.3525634Z * [new branch] gh/anijain2305/805/orig -> origin/gh/anijain2305/805/orig 2025-07-17T08:34:50.3530270Z * [new branch] gh/anijain2305/806/base -> origin/gh/anijain2305/806/base 2025-07-17T08:34:50.3533626Z * [new branch] gh/anijain2305/806/head -> origin/gh/anijain2305/806/head 2025-07-17T08:34:50.3537005Z * [new branch] gh/anijain2305/806/orig -> origin/gh/anijain2305/806/orig 2025-07-17T08:34:50.3541630Z * [new branch] gh/anijain2305/807/base -> origin/gh/anijain2305/807/base 2025-07-17T08:34:50.3545003Z * [new branch] gh/anijain2305/807/head -> origin/gh/anijain2305/807/head 2025-07-17T08:34:50.3548381Z * [new branch] gh/anijain2305/807/orig -> origin/gh/anijain2305/807/orig 2025-07-17T08:34:50.3552999Z * [new branch] gh/anijain2305/808/base -> origin/gh/anijain2305/808/base 2025-07-17T08:34:50.3556449Z * [new branch] gh/anijain2305/808/head -> origin/gh/anijain2305/808/head 2025-07-17T08:34:50.3559728Z * [new branch] gh/anijain2305/808/orig -> origin/gh/anijain2305/808/orig 2025-07-17T08:34:50.3564397Z * [new branch] gh/anijain2305/809/base -> origin/gh/anijain2305/809/base 2025-07-17T08:34:50.3567718Z * [new branch] gh/anijain2305/809/head -> origin/gh/anijain2305/809/head 2025-07-17T08:34:50.3571066Z * [new branch] gh/anijain2305/809/orig -> origin/gh/anijain2305/809/orig 2025-07-17T08:34:50.3575470Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-07-17T08:34:50.3578832Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-07-17T08:34:50.3582228Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-07-17T08:34:50.3586892Z * [new branch] gh/anijain2305/811/base -> origin/gh/anijain2305/811/base 2025-07-17T08:34:50.3590435Z * [new branch] gh/anijain2305/811/head -> origin/gh/anijain2305/811/head 2025-07-17T08:34:50.3593740Z * [new branch] gh/anijain2305/811/orig -> origin/gh/anijain2305/811/orig 2025-07-17T08:34:50.3598615Z * [new branch] gh/anijain2305/812/base -> origin/gh/anijain2305/812/base 2025-07-17T08:34:50.3601866Z * [new branch] gh/anijain2305/812/head -> origin/gh/anijain2305/812/head 2025-07-17T08:34:50.3605260Z * [new branch] gh/anijain2305/812/orig -> origin/gh/anijain2305/812/orig 2025-07-17T08:34:50.3609866Z * [new branch] gh/anijain2305/813/base -> origin/gh/anijain2305/813/base 2025-07-17T08:34:50.3613301Z * [new branch] gh/anijain2305/813/head -> origin/gh/anijain2305/813/head 2025-07-17T08:34:50.3616614Z * [new branch] gh/anijain2305/813/orig -> origin/gh/anijain2305/813/orig 2025-07-17T08:34:50.3621175Z * [new branch] gh/anijain2305/814/base -> origin/gh/anijain2305/814/base 2025-07-17T08:34:50.3624520Z * [new branch] gh/anijain2305/814/head -> origin/gh/anijain2305/814/head 2025-07-17T08:34:50.3628097Z * [new branch] gh/anijain2305/814/orig -> origin/gh/anijain2305/814/orig 2025-07-17T08:34:50.3632525Z * [new branch] gh/anijain2305/815/base -> origin/gh/anijain2305/815/base 2025-07-17T08:34:50.3637503Z * [new branch] gh/anijain2305/815/head -> origin/gh/anijain2305/815/head 2025-07-17T08:34:50.3640923Z * [new branch] gh/anijain2305/815/orig -> origin/gh/anijain2305/815/orig 2025-07-17T08:34:50.3645568Z * [new branch] gh/anijain2305/816/base -> origin/gh/anijain2305/816/base 2025-07-17T08:34:50.3648891Z * [new branch] gh/anijain2305/816/head -> origin/gh/anijain2305/816/head 2025-07-17T08:34:50.3653521Z * [new branch] gh/anijain2305/817/base -> origin/gh/anijain2305/817/base 2025-07-17T08:34:50.3657007Z * [new branch] gh/anijain2305/817/head -> origin/gh/anijain2305/817/head 2025-07-17T08:34:50.3660426Z * [new branch] gh/anijain2305/817/orig -> origin/gh/anijain2305/817/orig 2025-07-17T08:34:50.3665113Z * [new branch] gh/anijain2305/818/base -> origin/gh/anijain2305/818/base 2025-07-17T08:34:50.3668552Z * [new branch] gh/anijain2305/818/head -> origin/gh/anijain2305/818/head 2025-07-17T08:34:50.3671969Z * [new branch] gh/anijain2305/818/orig -> origin/gh/anijain2305/818/orig 2025-07-17T08:34:50.3676749Z * [new branch] gh/anijain2305/819/base -> origin/gh/anijain2305/819/base 2025-07-17T08:34:50.3680295Z * [new branch] gh/anijain2305/819/head -> origin/gh/anijain2305/819/head 2025-07-17T08:34:50.3683702Z * [new branch] gh/anijain2305/819/orig -> origin/gh/anijain2305/819/orig 2025-07-17T08:34:50.3688143Z * [new branch] gh/anijain2305/820/base -> origin/gh/anijain2305/820/base 2025-07-17T08:34:50.3691555Z * [new branch] gh/anijain2305/820/head -> origin/gh/anijain2305/820/head 2025-07-17T08:34:50.3694923Z * [new branch] gh/anijain2305/820/orig -> origin/gh/anijain2305/820/orig 2025-07-17T08:34:50.3700577Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-07-17T08:34:50.3703910Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-07-17T08:34:50.3707279Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-07-17T08:34:50.3712999Z * [new branch] gh/ankitageorge/10/base -> origin/gh/ankitageorge/10/base 2025-07-17T08:34:50.3716383Z * [new branch] gh/ankitageorge/10/head -> origin/gh/ankitageorge/10/head 2025-07-17T08:34:50.3719947Z * [new branch] gh/ankitageorge/10/orig -> origin/gh/ankitageorge/10/orig 2025-07-17T08:34:50.3724754Z * [new branch] gh/ankitageorge/11/base -> origin/gh/ankitageorge/11/base 2025-07-17T08:34:50.3728120Z * [new branch] gh/ankitageorge/11/head -> origin/gh/ankitageorge/11/head 2025-07-17T08:34:50.3731580Z * [new branch] gh/ankitageorge/11/orig -> origin/gh/ankitageorge/11/orig 2025-07-17T08:34:50.3736156Z * [new branch] gh/ankitageorge/12/base -> origin/gh/ankitageorge/12/base 2025-07-17T08:34:50.3739515Z * [new branch] gh/ankitageorge/12/head -> origin/gh/ankitageorge/12/head 2025-07-17T08:34:50.3742991Z * [new branch] gh/ankitageorge/12/orig -> origin/gh/ankitageorge/12/orig 2025-07-17T08:34:50.3747722Z * [new branch] gh/ankitageorge/5/base -> origin/gh/ankitageorge/5/base 2025-07-17T08:34:50.3751014Z * [new branch] gh/ankitageorge/5/head -> origin/gh/ankitageorge/5/head 2025-07-17T08:34:50.3756195Z * [new branch] gh/ankitageorge/5/orig -> origin/gh/ankitageorge/5/orig 2025-07-17T08:34:50.3760982Z * [new branch] gh/ankitageorge/7/base -> origin/gh/ankitageorge/7/base 2025-07-17T08:34:50.3764320Z * [new branch] gh/ankitageorge/7/head -> origin/gh/ankitageorge/7/head 2025-07-17T08:34:50.3768062Z * [new branch] gh/ankitageorge/7/orig -> origin/gh/ankitageorge/7/orig 2025-07-17T08:34:50.3772652Z * [new branch] gh/ankitageorge/8/base -> origin/gh/ankitageorge/8/base 2025-07-17T08:34:50.3775953Z * [new branch] gh/ankitageorge/8/head -> origin/gh/ankitageorge/8/head 2025-07-17T08:34:50.3779415Z * [new branch] gh/ankitageorge/8/orig -> origin/gh/ankitageorge/8/orig 2025-07-17T08:34:50.3783968Z * [new branch] gh/ankitageorge/9/base -> origin/gh/ankitageorge/9/base 2025-07-17T08:34:50.3787578Z * [new branch] gh/ankitageorge/9/head -> origin/gh/ankitageorge/9/head 2025-07-17T08:34:50.3790965Z * [new branch] gh/ankitageorge/9/orig -> origin/gh/ankitageorge/9/orig 2025-07-17T08:34:50.3796868Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-07-17T08:34:50.3800217Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-07-17T08:34:50.3804720Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-07-17T08:34:50.3808064Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-07-17T08:34:50.3812688Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-07-17T08:34:50.3815934Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-07-17T08:34:50.3820392Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-07-17T08:34:50.3823712Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-07-17T08:34:50.3828146Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-07-17T08:34:50.3831479Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-07-17T08:34:50.3836079Z * [new branch] gh/anshul-si/6/base -> origin/gh/anshul-si/6/base 2025-07-17T08:34:50.3839569Z * [new branch] gh/anshul-si/6/head -> origin/gh/anshul-si/6/head 2025-07-17T08:34:50.3843181Z * [new branch] gh/anshul-si/6/orig -> origin/gh/anshul-si/6/orig 2025-07-17T08:34:50.3848693Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-07-17T08:34:50.3852307Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-07-17T08:34:50.3856683Z * [new branch] gh/aorenste/231/base -> origin/gh/aorenste/231/base 2025-07-17T08:34:50.3860142Z * [new branch] gh/aorenste/231/head -> origin/gh/aorenste/231/head 2025-07-17T08:34:50.3863463Z * [new branch] gh/aorenste/231/orig -> origin/gh/aorenste/231/orig 2025-07-17T08:34:50.3868171Z * [new branch] gh/aorenste/232/base -> origin/gh/aorenste/232/base 2025-07-17T08:34:50.3871551Z * [new branch] gh/aorenste/232/head -> origin/gh/aorenste/232/head 2025-07-17T08:34:50.3875026Z * [new branch] gh/aorenste/232/orig -> origin/gh/aorenste/232/orig 2025-07-17T08:34:50.3879618Z * [new branch] gh/aorenste/234/base -> origin/gh/aorenste/234/base 2025-07-17T08:34:50.3883024Z * [new branch] gh/aorenste/234/head -> origin/gh/aorenste/234/head 2025-07-17T08:34:50.3886376Z * [new branch] gh/aorenste/234/orig -> origin/gh/aorenste/234/orig 2025-07-17T08:34:50.3891095Z * [new branch] gh/aorenste/235/base -> origin/gh/aorenste/235/base 2025-07-17T08:34:50.3894465Z * [new branch] gh/aorenste/235/head -> origin/gh/aorenste/235/head 2025-07-17T08:34:50.3898023Z * [new branch] gh/aorenste/235/orig -> origin/gh/aorenste/235/orig 2025-07-17T08:34:50.3902725Z * [new branch] gh/aorenste/236/base -> origin/gh/aorenste/236/base 2025-07-17T08:34:50.3906293Z * [new branch] gh/aorenste/236/head -> origin/gh/aorenste/236/head 2025-07-17T08:34:50.3909608Z * [new branch] gh/aorenste/236/orig -> origin/gh/aorenste/236/orig 2025-07-17T08:34:50.3915269Z * [new branch] gh/bdhirsh/647/base -> origin/gh/bdhirsh/647/base 2025-07-17T08:34:50.3918674Z * [new branch] gh/bdhirsh/647/head -> origin/gh/bdhirsh/647/head 2025-07-17T08:34:50.3922035Z * [new branch] gh/bdhirsh/647/orig -> origin/gh/bdhirsh/647/orig 2025-07-17T08:34:50.3926704Z * [new branch] gh/bdhirsh/650/base -> origin/gh/bdhirsh/650/base 2025-07-17T08:34:50.3930256Z * [new branch] gh/bdhirsh/650/head -> origin/gh/bdhirsh/650/head 2025-07-17T08:34:50.3933583Z * [new branch] gh/bdhirsh/650/orig -> origin/gh/bdhirsh/650/orig 2025-07-17T08:34:50.3938268Z * [new branch] gh/bdhirsh/651/base -> origin/gh/bdhirsh/651/base 2025-07-17T08:34:50.3941565Z * [new branch] gh/bdhirsh/651/head -> origin/gh/bdhirsh/651/head 2025-07-17T08:34:50.3944979Z * [new branch] gh/bdhirsh/651/orig -> origin/gh/bdhirsh/651/orig 2025-07-17T08:34:50.3949383Z * [new branch] gh/bdhirsh/652/base -> origin/gh/bdhirsh/652/base 2025-07-17T08:34:50.3952889Z * [new branch] gh/bdhirsh/652/head -> origin/gh/bdhirsh/652/head 2025-07-17T08:34:50.3956417Z * [new branch] gh/bdhirsh/652/orig -> origin/gh/bdhirsh/652/orig 2025-07-17T08:34:50.3960898Z * [new branch] gh/bdhirsh/655/base -> origin/gh/bdhirsh/655/base 2025-07-17T08:34:50.3964167Z * [new branch] gh/bdhirsh/655/head -> origin/gh/bdhirsh/655/head 2025-07-17T08:34:50.3967566Z * [new branch] gh/bdhirsh/655/orig -> origin/gh/bdhirsh/655/orig 2025-07-17T08:34:50.3972112Z * [new branch] gh/bdhirsh/656/base -> origin/gh/bdhirsh/656/base 2025-07-17T08:34:50.3975465Z * [new branch] gh/bdhirsh/656/head -> origin/gh/bdhirsh/656/head 2025-07-17T08:34:50.3980355Z * [new branch] gh/bdhirsh/657/base -> origin/gh/bdhirsh/657/base 2025-07-17T08:34:50.3983309Z * [new branch] gh/bdhirsh/657/head -> origin/gh/bdhirsh/657/head 2025-07-17T08:34:50.3988001Z * [new branch] gh/bdhirsh/659/base -> origin/gh/bdhirsh/659/base 2025-07-17T08:34:50.3991341Z * [new branch] gh/bdhirsh/659/head -> origin/gh/bdhirsh/659/head 2025-07-17T08:34:50.3997066Z * [new branch] gh/bdhirsh/659/orig -> origin/gh/bdhirsh/659/orig 2025-07-17T08:34:50.4001690Z * [new branch] gh/bdhirsh/663/base -> origin/gh/bdhirsh/663/base 2025-07-17T08:34:50.4005037Z * [new branch] gh/bdhirsh/663/head -> origin/gh/bdhirsh/663/head 2025-07-17T08:34:50.4008446Z * [new branch] gh/bdhirsh/663/orig -> origin/gh/bdhirsh/663/orig 2025-07-17T08:34:50.4013031Z * [new branch] gh/bdhirsh/665/base -> origin/gh/bdhirsh/665/base 2025-07-17T08:34:50.4016338Z * [new branch] gh/bdhirsh/665/head -> origin/gh/bdhirsh/665/head 2025-07-17T08:34:50.4019672Z * [new branch] gh/bdhirsh/665/orig -> origin/gh/bdhirsh/665/orig 2025-07-17T08:34:50.4024194Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-07-17T08:34:50.4027562Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-07-17T08:34:50.4030923Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-07-17T08:34:50.4036684Z * [new branch] gh/benjaminglass1/77/base -> origin/gh/benjaminglass1/77/base 2025-07-17T08:34:50.4039990Z * [new branch] gh/benjaminglass1/77/head -> origin/gh/benjaminglass1/77/head 2025-07-17T08:34:50.4043504Z * [new branch] gh/benjaminglass1/77/orig -> origin/gh/benjaminglass1/77/orig 2025-07-17T08:34:50.4047904Z * [new branch] gh/benjaminglass1/79/base -> origin/gh/benjaminglass1/79/base 2025-07-17T08:34:50.4051196Z * [new branch] gh/benjaminglass1/79/head -> origin/gh/benjaminglass1/79/head 2025-07-17T08:34:50.4054550Z * [new branch] gh/benjaminglass1/79/orig -> origin/gh/benjaminglass1/79/orig 2025-07-17T08:34:50.4059115Z * [new branch] gh/benjaminglass1/83/base -> origin/gh/benjaminglass1/83/base 2025-07-17T08:34:50.4062450Z * [new branch] gh/benjaminglass1/83/head -> origin/gh/benjaminglass1/83/head 2025-07-17T08:34:50.4065824Z * [new branch] gh/benjaminglass1/83/orig -> origin/gh/benjaminglass1/83/orig 2025-07-17T08:34:50.4070542Z * [new branch] gh/benjaminglass1/84/base -> origin/gh/benjaminglass1/84/base 2025-07-17T08:34:50.4074114Z * [new branch] gh/benjaminglass1/84/head -> origin/gh/benjaminglass1/84/head 2025-07-17T08:34:50.4077658Z * [new branch] gh/benjaminglass1/84/orig -> origin/gh/benjaminglass1/84/orig 2025-07-17T08:34:50.4082387Z * [new branch] gh/benjaminglass1/85/base -> origin/gh/benjaminglass1/85/base 2025-07-17T08:34:50.4085870Z * [new branch] gh/benjaminglass1/85/head -> origin/gh/benjaminglass1/85/head 2025-07-17T08:34:50.4089259Z * [new branch] gh/benjaminglass1/85/orig -> origin/gh/benjaminglass1/85/orig 2025-07-17T08:34:50.4093846Z * [new branch] gh/benjaminglass1/86/base -> origin/gh/benjaminglass1/86/base 2025-07-17T08:34:50.4097268Z * [new branch] gh/benjaminglass1/86/head -> origin/gh/benjaminglass1/86/head 2025-07-17T08:34:50.4100607Z * [new branch] gh/benjaminglass1/86/orig -> origin/gh/benjaminglass1/86/orig 2025-07-17T08:34:50.4105184Z * [new branch] gh/benjaminglass1/87/base -> origin/gh/benjaminglass1/87/base 2025-07-17T08:34:50.4108553Z * [new branch] gh/benjaminglass1/87/head -> origin/gh/benjaminglass1/87/head 2025-07-17T08:34:50.4111918Z * [new branch] gh/benjaminglass1/87/orig -> origin/gh/benjaminglass1/87/orig 2025-07-17T08:34:50.4116827Z * [new branch] gh/benjaminglass1/89/base -> origin/gh/benjaminglass1/89/base 2025-07-17T08:34:50.4120161Z * [new branch] gh/benjaminglass1/89/head -> origin/gh/benjaminglass1/89/head 2025-07-17T08:34:50.4123570Z * [new branch] gh/benjaminglass1/89/orig -> origin/gh/benjaminglass1/89/orig 2025-07-17T08:34:50.4128254Z * [new branch] gh/benjaminglass1/90/base -> origin/gh/benjaminglass1/90/base 2025-07-17T08:34:50.4131578Z * [new branch] gh/benjaminglass1/90/head -> origin/gh/benjaminglass1/90/head 2025-07-17T08:34:50.4135005Z * [new branch] gh/benjaminglass1/90/orig -> origin/gh/benjaminglass1/90/orig 2025-07-17T08:34:50.4139550Z * [new branch] gh/benjaminglass1/91/base -> origin/gh/benjaminglass1/91/base 2025-07-17T08:34:50.4142904Z * [new branch] gh/benjaminglass1/91/head -> origin/gh/benjaminglass1/91/head 2025-07-17T08:34:50.4146357Z * [new branch] gh/benjaminglass1/91/orig -> origin/gh/benjaminglass1/91/orig 2025-07-17T08:34:50.4150927Z * [new branch] gh/benjaminglass1/92/base -> origin/gh/benjaminglass1/92/base 2025-07-17T08:34:50.4154341Z * [new branch] gh/benjaminglass1/92/head -> origin/gh/benjaminglass1/92/head 2025-07-17T08:34:50.4157894Z * [new branch] gh/benjaminglass1/92/orig -> origin/gh/benjaminglass1/92/orig 2025-07-17T08:34:50.4162546Z * [new branch] gh/benjaminglass1/93/base -> origin/gh/benjaminglass1/93/base 2025-07-17T08:34:50.4165933Z * [new branch] gh/benjaminglass1/93/head -> origin/gh/benjaminglass1/93/head 2025-07-17T08:34:50.4169446Z * [new branch] gh/benjaminglass1/93/orig -> origin/gh/benjaminglass1/93/orig 2025-07-17T08:34:50.4174431Z * [new branch] gh/benjaminglass1/94/base -> origin/gh/benjaminglass1/94/base 2025-07-17T08:34:50.4177564Z * [new branch] gh/benjaminglass1/94/head -> origin/gh/benjaminglass1/94/head 2025-07-17T08:34:50.4181093Z * [new branch] gh/benjaminglass1/94/orig -> origin/gh/benjaminglass1/94/orig 2025-07-17T08:34:50.4186535Z * [new branch] gh/bobrenjc93/439/base -> origin/gh/bobrenjc93/439/base 2025-07-17T08:34:50.4191077Z * [new branch] gh/bobrenjc93/439/head -> origin/gh/bobrenjc93/439/head 2025-07-17T08:34:50.4195312Z * [new branch] gh/bobrenjc93/439/orig -> origin/gh/bobrenjc93/439/orig 2025-07-17T08:34:50.4199984Z * [new branch] gh/bobrenjc93/442/base -> origin/gh/bobrenjc93/442/base 2025-07-17T08:34:50.4203287Z * [new branch] gh/bobrenjc93/442/head -> origin/gh/bobrenjc93/442/head 2025-07-17T08:34:50.4206805Z * [new branch] gh/bobrenjc93/442/orig -> origin/gh/bobrenjc93/442/orig 2025-07-17T08:34:50.4211315Z * [new branch] gh/bobrenjc93/452/base -> origin/gh/bobrenjc93/452/base 2025-07-17T08:34:50.4214624Z * [new branch] gh/bobrenjc93/452/head -> origin/gh/bobrenjc93/452/head 2025-07-17T08:34:50.4218097Z * [new branch] gh/bobrenjc93/452/orig -> origin/gh/bobrenjc93/452/orig 2025-07-17T08:34:50.4222648Z * [new branch] gh/bobrenjc93/453/base -> origin/gh/bobrenjc93/453/base 2025-07-17T08:34:50.4226009Z * [new branch] gh/bobrenjc93/453/head -> origin/gh/bobrenjc93/453/head 2025-07-17T08:34:50.4229296Z * [new branch] gh/bobrenjc93/453/orig -> origin/gh/bobrenjc93/453/orig 2025-07-17T08:34:50.4233776Z * [new branch] gh/bobrenjc93/454/base -> origin/gh/bobrenjc93/454/base 2025-07-17T08:34:50.4238419Z * [new branch] gh/bobrenjc93/454/head -> origin/gh/bobrenjc93/454/head 2025-07-17T08:34:50.4241791Z * [new branch] gh/bobrenjc93/454/orig -> origin/gh/bobrenjc93/454/orig 2025-07-17T08:34:50.4246369Z * [new branch] gh/bobrenjc93/455/base -> origin/gh/bobrenjc93/455/base 2025-07-17T08:34:50.4249758Z * [new branch] gh/bobrenjc93/455/head -> origin/gh/bobrenjc93/455/head 2025-07-17T08:34:50.4253166Z * [new branch] gh/bobrenjc93/455/orig -> origin/gh/bobrenjc93/455/orig 2025-07-17T08:34:50.4257737Z * [new branch] gh/bobrenjc93/456/base -> origin/gh/bobrenjc93/456/base 2025-07-17T08:34:50.4261058Z * [new branch] gh/bobrenjc93/456/head -> origin/gh/bobrenjc93/456/head 2025-07-17T08:34:50.4264580Z * [new branch] gh/bobrenjc93/456/orig -> origin/gh/bobrenjc93/456/orig 2025-07-17T08:34:50.4269298Z * [new branch] gh/bobrenjc93/457/base -> origin/gh/bobrenjc93/457/base 2025-07-17T08:34:50.4272636Z * [new branch] gh/bobrenjc93/457/head -> origin/gh/bobrenjc93/457/head 2025-07-17T08:34:50.4276209Z * [new branch] gh/bobrenjc93/457/orig -> origin/gh/bobrenjc93/457/orig 2025-07-17T08:34:50.4281108Z * [new branch] gh/bobrenjc93/458/base -> origin/gh/bobrenjc93/458/base 2025-07-17T08:34:50.4284045Z * [new branch] gh/bobrenjc93/458/head -> origin/gh/bobrenjc93/458/head 2025-07-17T08:34:50.4287459Z * [new branch] gh/bobrenjc93/458/orig -> origin/gh/bobrenjc93/458/orig 2025-07-17T08:34:50.4292138Z * [new branch] gh/bobrenjc93/459/base -> origin/gh/bobrenjc93/459/base 2025-07-17T08:34:50.4295487Z * [new branch] gh/bobrenjc93/459/head -> origin/gh/bobrenjc93/459/head 2025-07-17T08:34:50.4298882Z * [new branch] gh/bobrenjc93/459/orig -> origin/gh/bobrenjc93/459/orig 2025-07-17T08:34:50.4303372Z * [new branch] gh/bobrenjc93/460/base -> origin/gh/bobrenjc93/460/base 2025-07-17T08:34:50.4306954Z * [new branch] gh/bobrenjc93/460/head -> origin/gh/bobrenjc93/460/head 2025-07-17T08:34:50.4310297Z * [new branch] gh/bobrenjc93/460/orig -> origin/gh/bobrenjc93/460/orig 2025-07-17T08:34:50.4315012Z * [new branch] gh/bobrenjc93/461/base -> origin/gh/bobrenjc93/461/base 2025-07-17T08:34:50.4318436Z * [new branch] gh/bobrenjc93/461/head -> origin/gh/bobrenjc93/461/head 2025-07-17T08:34:50.4321847Z * [new branch] gh/bobrenjc93/461/orig -> origin/gh/bobrenjc93/461/orig 2025-07-17T08:34:50.4326346Z * [new branch] gh/bobrenjc93/462/base -> origin/gh/bobrenjc93/462/base 2025-07-17T08:34:50.4329679Z * [new branch] gh/bobrenjc93/462/head -> origin/gh/bobrenjc93/462/head 2025-07-17T08:34:50.4333110Z * [new branch] gh/bobrenjc93/462/orig -> origin/gh/bobrenjc93/462/orig 2025-07-17T08:34:50.4337673Z * [new branch] gh/bobrenjc93/463/base -> origin/gh/bobrenjc93/463/base 2025-07-17T08:34:50.4340980Z * [new branch] gh/bobrenjc93/463/head -> origin/gh/bobrenjc93/463/head 2025-07-17T08:34:50.4344403Z * [new branch] gh/bobrenjc93/463/orig -> origin/gh/bobrenjc93/463/orig 2025-07-17T08:34:50.4349182Z * [new branch] gh/bobrenjc93/464/base -> origin/gh/bobrenjc93/464/base 2025-07-17T08:34:50.4352341Z * [new branch] gh/bobrenjc93/464/head -> origin/gh/bobrenjc93/464/head 2025-07-17T08:34:50.4356326Z * [new branch] gh/bobrenjc93/464/orig -> origin/gh/bobrenjc93/464/orig 2025-07-17T08:34:50.4361105Z * [new branch] gh/bobrenjc93/465/base -> origin/gh/bobrenjc93/465/base 2025-07-17T08:34:50.4364442Z * [new branch] gh/bobrenjc93/465/head -> origin/gh/bobrenjc93/465/head 2025-07-17T08:34:50.4367777Z * [new branch] gh/bobrenjc93/465/orig -> origin/gh/bobrenjc93/465/orig 2025-07-17T08:34:50.4372529Z * [new branch] gh/bobrenjc93/466/base -> origin/gh/bobrenjc93/466/base 2025-07-17T08:34:50.4375899Z * [new branch] gh/bobrenjc93/466/head -> origin/gh/bobrenjc93/466/head 2025-07-17T08:34:50.4379253Z * [new branch] gh/bobrenjc93/466/orig -> origin/gh/bobrenjc93/466/orig 2025-07-17T08:34:50.4383829Z * [new branch] gh/bobrenjc93/467/base -> origin/gh/bobrenjc93/467/base 2025-07-17T08:34:50.4387222Z * [new branch] gh/bobrenjc93/467/head -> origin/gh/bobrenjc93/467/head 2025-07-17T08:34:50.4390725Z * [new branch] gh/bobrenjc93/467/orig -> origin/gh/bobrenjc93/467/orig 2025-07-17T08:34:50.4395599Z * [new branch] gh/bobrenjc93/468/base -> origin/gh/bobrenjc93/468/base 2025-07-17T08:34:50.4398942Z * [new branch] gh/bobrenjc93/468/head -> origin/gh/bobrenjc93/468/head 2025-07-17T08:34:50.4402378Z * [new branch] gh/bobrenjc93/468/orig -> origin/gh/bobrenjc93/468/orig 2025-07-17T08:34:50.4407014Z * [new branch] gh/bobrenjc93/469/base -> origin/gh/bobrenjc93/469/base 2025-07-17T08:34:50.4410402Z * [new branch] gh/bobrenjc93/469/head -> origin/gh/bobrenjc93/469/head 2025-07-17T08:34:50.4413796Z * [new branch] gh/bobrenjc93/469/orig -> origin/gh/bobrenjc93/469/orig 2025-07-17T08:34:50.4418329Z * [new branch] gh/bobrenjc93/470/base -> origin/gh/bobrenjc93/470/base 2025-07-17T08:34:50.4421662Z * [new branch] gh/bobrenjc93/470/head -> origin/gh/bobrenjc93/470/head 2025-07-17T08:34:50.4425086Z * [new branch] gh/bobrenjc93/470/orig -> origin/gh/bobrenjc93/470/orig 2025-07-17T08:34:50.4429677Z * [new branch] gh/bobrenjc93/471/base -> origin/gh/bobrenjc93/471/base 2025-07-17T08:34:50.4433056Z * [new branch] gh/bobrenjc93/471/head -> origin/gh/bobrenjc93/471/head 2025-07-17T08:34:50.4436976Z * [new branch] gh/bobrenjc93/471/orig -> origin/gh/bobrenjc93/471/orig 2025-07-17T08:34:50.4441356Z * [new branch] gh/bobrenjc93/472/base -> origin/gh/bobrenjc93/472/base 2025-07-17T08:34:50.4444761Z * [new branch] gh/bobrenjc93/472/head -> origin/gh/bobrenjc93/472/head 2025-07-17T08:34:50.4448172Z * [new branch] gh/bobrenjc93/472/orig -> origin/gh/bobrenjc93/472/orig 2025-07-17T08:34:50.4452932Z * [new branch] gh/bobrenjc93/473/base -> origin/gh/bobrenjc93/473/base 2025-07-17T08:34:50.4456326Z * [new branch] gh/bobrenjc93/473/head -> origin/gh/bobrenjc93/473/head 2025-07-17T08:34:50.4459700Z * [new branch] gh/bobrenjc93/473/orig -> origin/gh/bobrenjc93/473/orig 2025-07-17T08:34:50.4464295Z * [new branch] gh/bobrenjc93/474/base -> origin/gh/bobrenjc93/474/base 2025-07-17T08:34:50.4467687Z * [new branch] gh/bobrenjc93/474/head -> origin/gh/bobrenjc93/474/head 2025-07-17T08:34:50.4471054Z * [new branch] gh/bobrenjc93/474/orig -> origin/gh/bobrenjc93/474/orig 2025-07-17T08:34:50.4495548Z * [new branch] gh/bobrenjc93/475/base -> origin/gh/bobrenjc93/475/base 2025-07-17T08:34:50.4495955Z * [new branch] gh/bobrenjc93/475/head -> origin/gh/bobrenjc93/475/head 2025-07-17T08:34:50.4496331Z * [new branch] gh/bobrenjc93/475/orig -> origin/gh/bobrenjc93/475/orig 2025-07-17T08:34:50.4496671Z * [new branch] gh/bobrenjc93/476/base -> origin/gh/bobrenjc93/476/base 2025-07-17T08:34:50.4497018Z * [new branch] gh/bobrenjc93/476/head -> origin/gh/bobrenjc93/476/head 2025-07-17T08:34:50.4497419Z * [new branch] gh/bobrenjc93/476/orig -> origin/gh/bobrenjc93/476/orig 2025-07-17T08:34:50.4500368Z * [new branch] gh/bobrenjc93/477/base -> origin/gh/bobrenjc93/477/base 2025-07-17T08:34:50.4503694Z * [new branch] gh/bobrenjc93/477/head -> origin/gh/bobrenjc93/477/head 2025-07-17T08:34:50.4507034Z * [new branch] gh/bobrenjc93/477/orig -> origin/gh/bobrenjc93/477/orig 2025-07-17T08:34:50.4511642Z * [new branch] gh/bobrenjc93/478/base -> origin/gh/bobrenjc93/478/base 2025-07-17T08:34:50.4515184Z * [new branch] gh/bobrenjc93/478/head -> origin/gh/bobrenjc93/478/head 2025-07-17T08:34:50.4518725Z * [new branch] gh/bobrenjc93/478/orig -> origin/gh/bobrenjc93/478/orig 2025-07-17T08:34:50.4523261Z * [new branch] gh/bobrenjc93/479/base -> origin/gh/bobrenjc93/479/base 2025-07-17T08:34:50.4526574Z * [new branch] gh/bobrenjc93/479/head -> origin/gh/bobrenjc93/479/head 2025-07-17T08:34:50.4529978Z * [new branch] gh/bobrenjc93/479/orig -> origin/gh/bobrenjc93/479/orig 2025-07-17T08:34:50.4534483Z * [new branch] gh/bobrenjc93/480/base -> origin/gh/bobrenjc93/480/base 2025-07-17T08:34:50.4537789Z * [new branch] gh/bobrenjc93/480/head -> origin/gh/bobrenjc93/480/head 2025-07-17T08:34:50.4541231Z * [new branch] gh/bobrenjc93/480/orig -> origin/gh/bobrenjc93/480/orig 2025-07-17T08:34:50.4545855Z * [new branch] gh/bobrenjc93/481/base -> origin/gh/bobrenjc93/481/base 2025-07-17T08:34:50.4549259Z * [new branch] gh/bobrenjc93/481/head -> origin/gh/bobrenjc93/481/head 2025-07-17T08:34:50.4552633Z * [new branch] gh/bobrenjc93/481/orig -> origin/gh/bobrenjc93/481/orig 2025-07-17T08:34:50.4557354Z * [new branch] gh/bobrenjc93/482/base -> origin/gh/bobrenjc93/482/base 2025-07-17T08:34:50.4560597Z * [new branch] gh/bobrenjc93/482/head -> origin/gh/bobrenjc93/482/head 2025-07-17T08:34:50.4564180Z * [new branch] gh/bobrenjc93/482/orig -> origin/gh/bobrenjc93/482/orig 2025-07-17T08:34:50.4568692Z * [new branch] gh/bobrenjc93/483/base -> origin/gh/bobrenjc93/483/base 2025-07-17T08:34:50.4572438Z * [new branch] gh/bobrenjc93/483/head -> origin/gh/bobrenjc93/483/head 2025-07-17T08:34:50.4575348Z * [new branch] gh/bobrenjc93/483/orig -> origin/gh/bobrenjc93/483/orig 2025-07-17T08:34:50.4579821Z * [new branch] gh/bobrenjc93/484/base -> origin/gh/bobrenjc93/484/base 2025-07-17T08:34:50.4583147Z * [new branch] gh/bobrenjc93/484/head -> origin/gh/bobrenjc93/484/head 2025-07-17T08:34:50.4586541Z * [new branch] gh/bobrenjc93/484/orig -> origin/gh/bobrenjc93/484/orig 2025-07-17T08:34:50.4591134Z * [new branch] gh/bobrenjc93/485/base -> origin/gh/bobrenjc93/485/base 2025-07-17T08:34:50.4594576Z * [new branch] gh/bobrenjc93/485/head -> origin/gh/bobrenjc93/485/head 2025-07-17T08:34:50.4598023Z * [new branch] gh/bobrenjc93/485/orig -> origin/gh/bobrenjc93/485/orig 2025-07-17T08:34:50.4602468Z * [new branch] gh/bobrenjc93/486/base -> origin/gh/bobrenjc93/486/base 2025-07-17T08:34:50.4605775Z * [new branch] gh/bobrenjc93/486/head -> origin/gh/bobrenjc93/486/head 2025-07-17T08:34:50.4609063Z * [new branch] gh/bobrenjc93/486/orig -> origin/gh/bobrenjc93/486/orig 2025-07-17T08:34:50.4613523Z * [new branch] gh/bobrenjc93/487/base -> origin/gh/bobrenjc93/487/base 2025-07-17T08:34:50.4616827Z * [new branch] gh/bobrenjc93/487/head -> origin/gh/bobrenjc93/487/head 2025-07-17T08:34:50.4620126Z * [new branch] gh/bobrenjc93/487/orig -> origin/gh/bobrenjc93/487/orig 2025-07-17T08:34:50.4624615Z * [new branch] gh/bobrenjc93/488/base -> origin/gh/bobrenjc93/488/base 2025-07-17T08:34:50.4627923Z * [new branch] gh/bobrenjc93/488/head -> origin/gh/bobrenjc93/488/head 2025-07-17T08:34:50.4631301Z * [new branch] gh/bobrenjc93/488/orig -> origin/gh/bobrenjc93/488/orig 2025-07-17T08:34:50.4636284Z * [new branch] gh/bobrenjc93/489/base -> origin/gh/bobrenjc93/489/base 2025-07-17T08:34:50.4639517Z * [new branch] gh/bobrenjc93/489/head -> origin/gh/bobrenjc93/489/head 2025-07-17T08:34:50.4643055Z * [new branch] gh/bobrenjc93/489/orig -> origin/gh/bobrenjc93/489/orig 2025-07-17T08:34:50.4647581Z * [new branch] gh/bobrenjc93/490/base -> origin/gh/bobrenjc93/490/base 2025-07-17T08:34:50.4650888Z * [new branch] gh/bobrenjc93/490/head -> origin/gh/bobrenjc93/490/head 2025-07-17T08:34:50.4654221Z * [new branch] gh/bobrenjc93/490/orig -> origin/gh/bobrenjc93/490/orig 2025-07-17T08:34:50.4658793Z * [new branch] gh/bobrenjc93/491/base -> origin/gh/bobrenjc93/491/base 2025-07-17T08:34:50.4662086Z * [new branch] gh/bobrenjc93/491/head -> origin/gh/bobrenjc93/491/head 2025-07-17T08:34:50.4665897Z * [new branch] gh/bobrenjc93/491/orig -> origin/gh/bobrenjc93/491/orig 2025-07-17T08:34:50.4670473Z * [new branch] gh/bobrenjc93/492/base -> origin/gh/bobrenjc93/492/base 2025-07-17T08:34:50.4674005Z * [new branch] gh/bobrenjc93/492/head -> origin/gh/bobrenjc93/492/head 2025-07-17T08:34:50.4677887Z * [new branch] gh/bobrenjc93/492/orig -> origin/gh/bobrenjc93/492/orig 2025-07-17T08:34:50.4682387Z * [new branch] gh/bobrenjc93/493/base -> origin/gh/bobrenjc93/493/base 2025-07-17T08:34:50.4685758Z * [new branch] gh/bobrenjc93/493/head -> origin/gh/bobrenjc93/493/head 2025-07-17T08:34:50.4689108Z * [new branch] gh/bobrenjc93/493/orig -> origin/gh/bobrenjc93/493/orig 2025-07-17T08:34:50.4693604Z * [new branch] gh/bobrenjc93/494/base -> origin/gh/bobrenjc93/494/base 2025-07-17T08:34:50.4696942Z * [new branch] gh/bobrenjc93/494/head -> origin/gh/bobrenjc93/494/head 2025-07-17T08:34:50.4700461Z * [new branch] gh/bobrenjc93/494/orig -> origin/gh/bobrenjc93/494/orig 2025-07-17T08:34:50.4704801Z * [new branch] gh/bobrenjc93/495/base -> origin/gh/bobrenjc93/495/base 2025-07-17T08:34:50.4708078Z * [new branch] gh/bobrenjc93/495/head -> origin/gh/bobrenjc93/495/head 2025-07-17T08:34:50.4711441Z * [new branch] gh/bobrenjc93/495/orig -> origin/gh/bobrenjc93/495/orig 2025-07-17T08:34:50.4717884Z * [new branch] gh/bobrenjc93/496/base -> origin/gh/bobrenjc93/496/base 2025-07-17T08:34:50.4721223Z * [new branch] gh/bobrenjc93/496/head -> origin/gh/bobrenjc93/496/head 2025-07-17T08:34:50.4724722Z * [new branch] gh/bobrenjc93/496/orig -> origin/gh/bobrenjc93/496/orig 2025-07-17T08:34:50.4729386Z * [new branch] gh/bobrenjc93/497/base -> origin/gh/bobrenjc93/497/base 2025-07-17T08:34:50.4732682Z * [new branch] gh/bobrenjc93/497/head -> origin/gh/bobrenjc93/497/head 2025-07-17T08:34:50.4735977Z * [new branch] gh/bobrenjc93/497/orig -> origin/gh/bobrenjc93/497/orig 2025-07-17T08:34:50.4740468Z * [new branch] gh/bobrenjc93/498/base -> origin/gh/bobrenjc93/498/base 2025-07-17T08:34:50.4743772Z * [new branch] gh/bobrenjc93/498/head -> origin/gh/bobrenjc93/498/head 2025-07-17T08:34:50.4747110Z * [new branch] gh/bobrenjc93/498/orig -> origin/gh/bobrenjc93/498/orig 2025-07-17T08:34:50.4751625Z * [new branch] gh/bobrenjc93/499/base -> origin/gh/bobrenjc93/499/base 2025-07-17T08:34:50.4755310Z * [new branch] gh/bobrenjc93/499/head -> origin/gh/bobrenjc93/499/head 2025-07-17T08:34:50.4758666Z * [new branch] gh/bobrenjc93/499/orig -> origin/gh/bobrenjc93/499/orig 2025-07-17T08:34:50.4763165Z * [new branch] gh/bobrenjc93/500/base -> origin/gh/bobrenjc93/500/base 2025-07-17T08:34:50.4766490Z * [new branch] gh/bobrenjc93/500/head -> origin/gh/bobrenjc93/500/head 2025-07-17T08:34:50.4769960Z * [new branch] gh/bobrenjc93/500/orig -> origin/gh/bobrenjc93/500/orig 2025-07-17T08:34:50.4774453Z * [new branch] gh/bobrenjc93/501/base -> origin/gh/bobrenjc93/501/base 2025-07-17T08:34:50.4777792Z * [new branch] gh/bobrenjc93/501/head -> origin/gh/bobrenjc93/501/head 2025-07-17T08:34:50.4781147Z * [new branch] gh/bobrenjc93/501/orig -> origin/gh/bobrenjc93/501/orig 2025-07-17T08:34:50.4785659Z * [new branch] gh/bobrenjc93/502/base -> origin/gh/bobrenjc93/502/base 2025-07-17T08:34:50.4788982Z * [new branch] gh/bobrenjc93/502/head -> origin/gh/bobrenjc93/502/head 2025-07-17T08:34:50.4792279Z * [new branch] gh/bobrenjc93/502/orig -> origin/gh/bobrenjc93/502/orig 2025-07-17T08:34:50.4797207Z * [new branch] gh/bobrenjc93/503/base -> origin/gh/bobrenjc93/503/base 2025-07-17T08:34:50.4800475Z * [new branch] gh/bobrenjc93/503/head -> origin/gh/bobrenjc93/503/head 2025-07-17T08:34:50.4803796Z * [new branch] gh/bobrenjc93/503/orig -> origin/gh/bobrenjc93/503/orig 2025-07-17T08:34:50.4808459Z * [new branch] gh/bobrenjc93/504/base -> origin/gh/bobrenjc93/504/base 2025-07-17T08:34:50.4811760Z * [new branch] gh/bobrenjc93/504/head -> origin/gh/bobrenjc93/504/head 2025-07-17T08:34:50.4815204Z * [new branch] gh/bobrenjc93/504/orig -> origin/gh/bobrenjc93/504/orig 2025-07-17T08:34:50.4819925Z * [new branch] gh/bobrenjc93/505/base -> origin/gh/bobrenjc93/505/base 2025-07-17T08:34:50.4823153Z * [new branch] gh/bobrenjc93/505/head -> origin/gh/bobrenjc93/505/head 2025-07-17T08:34:50.4826523Z * [new branch] gh/bobrenjc93/505/orig -> origin/gh/bobrenjc93/505/orig 2025-07-17T08:34:50.4831209Z * [new branch] gh/bobrenjc93/506/base -> origin/gh/bobrenjc93/506/base 2025-07-17T08:34:50.4834538Z * [new branch] gh/bobrenjc93/506/head -> origin/gh/bobrenjc93/506/head 2025-07-17T08:34:50.4837938Z * [new branch] gh/bobrenjc93/506/orig -> origin/gh/bobrenjc93/506/orig 2025-07-17T08:34:50.4842404Z * [new branch] gh/bobrenjc93/507/base -> origin/gh/bobrenjc93/507/base 2025-07-17T08:34:50.4845793Z * [new branch] gh/bobrenjc93/507/head -> origin/gh/bobrenjc93/507/head 2025-07-17T08:34:50.4849070Z * [new branch] gh/bobrenjc93/507/orig -> origin/gh/bobrenjc93/507/orig 2025-07-17T08:34:50.4853581Z * [new branch] gh/bobrenjc93/508/base -> origin/gh/bobrenjc93/508/base 2025-07-17T08:34:50.4856991Z * [new branch] gh/bobrenjc93/508/head -> origin/gh/bobrenjc93/508/head 2025-07-17T08:34:50.4860419Z * [new branch] gh/bobrenjc93/508/orig -> origin/gh/bobrenjc93/508/orig 2025-07-17T08:34:50.4864987Z * [new branch] gh/bobrenjc93/509/base -> origin/gh/bobrenjc93/509/base 2025-07-17T08:34:50.4868326Z * [new branch] gh/bobrenjc93/509/head -> origin/gh/bobrenjc93/509/head 2025-07-17T08:34:50.4871648Z * [new branch] gh/bobrenjc93/509/orig -> origin/gh/bobrenjc93/509/orig 2025-07-17T08:34:50.4876382Z * [new branch] gh/bobrenjc93/510/base -> origin/gh/bobrenjc93/510/base 2025-07-17T08:34:50.4879677Z * [new branch] gh/bobrenjc93/510/head -> origin/gh/bobrenjc93/510/head 2025-07-17T08:34:50.4882990Z * [new branch] gh/bobrenjc93/510/orig -> origin/gh/bobrenjc93/510/orig 2025-07-17T08:34:50.4887540Z * [new branch] gh/bobrenjc93/511/base -> origin/gh/bobrenjc93/511/base 2025-07-17T08:34:50.4890855Z * [new branch] gh/bobrenjc93/511/head -> origin/gh/bobrenjc93/511/head 2025-07-17T08:34:50.4894160Z * [new branch] gh/bobrenjc93/511/orig -> origin/gh/bobrenjc93/511/orig 2025-07-17T08:34:50.4898654Z * [new branch] gh/bobrenjc93/512/base -> origin/gh/bobrenjc93/512/base 2025-07-17T08:34:50.4901996Z * [new branch] gh/bobrenjc93/512/head -> origin/gh/bobrenjc93/512/head 2025-07-17T08:34:50.4905429Z * [new branch] gh/bobrenjc93/512/orig -> origin/gh/bobrenjc93/512/orig 2025-07-17T08:34:50.4910133Z * [new branch] gh/bobrenjc93/513/base -> origin/gh/bobrenjc93/513/base 2025-07-17T08:34:50.4913484Z * [new branch] gh/bobrenjc93/513/head -> origin/gh/bobrenjc93/513/head 2025-07-17T08:34:50.4917043Z * [new branch] gh/bobrenjc93/513/orig -> origin/gh/bobrenjc93/513/orig 2025-07-17T08:34:50.4921478Z * [new branch] gh/bobrenjc93/514/base -> origin/gh/bobrenjc93/514/base 2025-07-17T08:34:50.4924825Z * [new branch] gh/bobrenjc93/514/head -> origin/gh/bobrenjc93/514/head 2025-07-17T08:34:50.4928202Z * [new branch] gh/bobrenjc93/514/orig -> origin/gh/bobrenjc93/514/orig 2025-07-17T08:34:50.4932798Z * [new branch] gh/bobrenjc93/515/base -> origin/gh/bobrenjc93/515/base 2025-07-17T08:34:50.4936151Z * [new branch] gh/bobrenjc93/515/head -> origin/gh/bobrenjc93/515/head 2025-07-17T08:34:50.4939423Z * [new branch] gh/bobrenjc93/515/orig -> origin/gh/bobrenjc93/515/orig 2025-07-17T08:34:50.4944112Z * [new branch] gh/bobrenjc93/516/base -> origin/gh/bobrenjc93/516/base 2025-07-17T08:34:50.4947422Z * [new branch] gh/bobrenjc93/516/head -> origin/gh/bobrenjc93/516/head 2025-07-17T08:34:50.4950910Z * [new branch] gh/bobrenjc93/516/orig -> origin/gh/bobrenjc93/516/orig 2025-07-17T08:34:50.4957367Z * [new branch] gh/bobrenjc93/517/base -> origin/gh/bobrenjc93/517/base 2025-07-17T08:34:50.4960628Z * [new branch] gh/bobrenjc93/517/head -> origin/gh/bobrenjc93/517/head 2025-07-17T08:34:50.4964126Z * [new branch] gh/bobrenjc93/517/orig -> origin/gh/bobrenjc93/517/orig 2025-07-17T08:34:50.4968438Z * [new branch] gh/bobrenjc93/518/base -> origin/gh/bobrenjc93/518/base 2025-07-17T08:34:50.4971777Z * [new branch] gh/bobrenjc93/518/head -> origin/gh/bobrenjc93/518/head 2025-07-17T08:34:50.4975041Z * [new branch] gh/bobrenjc93/518/orig -> origin/gh/bobrenjc93/518/orig 2025-07-17T08:34:50.4979485Z * [new branch] gh/bobrenjc93/519/base -> origin/gh/bobrenjc93/519/base 2025-07-17T08:34:50.4982853Z * [new branch] gh/bobrenjc93/519/head -> origin/gh/bobrenjc93/519/head 2025-07-17T08:34:50.4986242Z * [new branch] gh/bobrenjc93/519/orig -> origin/gh/bobrenjc93/519/orig 2025-07-17T08:34:50.4990750Z * [new branch] gh/bobrenjc93/520/base -> origin/gh/bobrenjc93/520/base 2025-07-17T08:34:50.4994179Z * [new branch] gh/bobrenjc93/520/head -> origin/gh/bobrenjc93/520/head 2025-07-17T08:34:50.4997712Z * [new branch] gh/bobrenjc93/520/orig -> origin/gh/bobrenjc93/520/orig 2025-07-17T08:34:50.5002368Z * [new branch] gh/bobrenjc93/521/base -> origin/gh/bobrenjc93/521/base 2025-07-17T08:34:50.5005724Z * [new branch] gh/bobrenjc93/521/head -> origin/gh/bobrenjc93/521/head 2025-07-17T08:34:50.5009007Z * [new branch] gh/bobrenjc93/521/orig -> origin/gh/bobrenjc93/521/orig 2025-07-17T08:34:50.5013572Z * [new branch] gh/bobrenjc93/522/base -> origin/gh/bobrenjc93/522/base 2025-07-17T08:34:50.5016863Z * [new branch] gh/bobrenjc93/522/head -> origin/gh/bobrenjc93/522/head 2025-07-17T08:34:50.5020210Z * [new branch] gh/bobrenjc93/522/orig -> origin/gh/bobrenjc93/522/orig 2025-07-17T08:34:50.5024777Z * [new branch] gh/bobrenjc93/523/base -> origin/gh/bobrenjc93/523/base 2025-07-17T08:34:50.5028130Z * [new branch] gh/bobrenjc93/523/head -> origin/gh/bobrenjc93/523/head 2025-07-17T08:34:50.5031424Z * [new branch] gh/bobrenjc93/523/orig -> origin/gh/bobrenjc93/523/orig 2025-07-17T08:34:50.5036178Z * [new branch] gh/bobrenjc93/524/base -> origin/gh/bobrenjc93/524/base 2025-07-17T08:34:50.5039425Z * [new branch] gh/bobrenjc93/524/head -> origin/gh/bobrenjc93/524/head 2025-07-17T08:34:50.5042921Z * [new branch] gh/bobrenjc93/524/orig -> origin/gh/bobrenjc93/524/orig 2025-07-17T08:34:50.5047571Z * [new branch] gh/bobrenjc93/525/base -> origin/gh/bobrenjc93/525/base 2025-07-17T08:34:50.5050846Z * [new branch] gh/bobrenjc93/525/head -> origin/gh/bobrenjc93/525/head 2025-07-17T08:34:50.5054219Z * [new branch] gh/bobrenjc93/525/orig -> origin/gh/bobrenjc93/525/orig 2025-07-17T08:34:50.5058910Z * [new branch] gh/bobrenjc93/526/base -> origin/gh/bobrenjc93/526/base 2025-07-17T08:34:50.5062053Z * [new branch] gh/bobrenjc93/526/head -> origin/gh/bobrenjc93/526/head 2025-07-17T08:34:50.5065480Z * [new branch] gh/bobrenjc93/526/orig -> origin/gh/bobrenjc93/526/orig 2025-07-17T08:34:50.5070115Z * [new branch] gh/bobrenjc93/527/base -> origin/gh/bobrenjc93/527/base 2025-07-17T08:34:50.5073396Z * [new branch] gh/bobrenjc93/527/head -> origin/gh/bobrenjc93/527/head 2025-07-17T08:34:50.5077012Z * [new branch] gh/bobrenjc93/527/orig -> origin/gh/bobrenjc93/527/orig 2025-07-17T08:34:50.5081449Z * [new branch] gh/bobrenjc93/528/base -> origin/gh/bobrenjc93/528/base 2025-07-17T08:34:50.5084766Z * [new branch] gh/bobrenjc93/528/head -> origin/gh/bobrenjc93/528/head 2025-07-17T08:34:50.5088263Z * [new branch] gh/bobrenjc93/528/orig -> origin/gh/bobrenjc93/528/orig 2025-07-17T08:34:50.5093149Z * [new branch] gh/bobrenjc93/529/base -> origin/gh/bobrenjc93/529/base 2025-07-17T08:34:50.5096272Z * [new branch] gh/bobrenjc93/529/head -> origin/gh/bobrenjc93/529/head 2025-07-17T08:34:50.5100047Z * [new branch] gh/bobrenjc93/529/orig -> origin/gh/bobrenjc93/529/orig 2025-07-17T08:34:50.5104112Z * [new branch] gh/bobrenjc93/530/base -> origin/gh/bobrenjc93/530/base 2025-07-17T08:34:50.5107486Z * [new branch] gh/bobrenjc93/530/head -> origin/gh/bobrenjc93/530/head 2025-07-17T08:34:50.5110863Z * [new branch] gh/bobrenjc93/530/orig -> origin/gh/bobrenjc93/530/orig 2025-07-17T08:34:50.5115627Z * [new branch] gh/bobrenjc93/531/base -> origin/gh/bobrenjc93/531/base 2025-07-17T08:34:50.5118898Z * [new branch] gh/bobrenjc93/531/head -> origin/gh/bobrenjc93/531/head 2025-07-17T08:34:50.5122159Z * [new branch] gh/bobrenjc93/531/orig -> origin/gh/bobrenjc93/531/orig 2025-07-17T08:34:50.5126799Z * [new branch] gh/bobrenjc93/532/base -> origin/gh/bobrenjc93/532/base 2025-07-17T08:34:50.5130041Z * [new branch] gh/bobrenjc93/532/head -> origin/gh/bobrenjc93/532/head 2025-07-17T08:34:50.5133668Z * [new branch] gh/bobrenjc93/532/orig -> origin/gh/bobrenjc93/532/orig 2025-07-17T08:34:50.5138313Z * [new branch] gh/bobrenjc93/533/base -> origin/gh/bobrenjc93/533/base 2025-07-17T08:34:50.5141692Z * [new branch] gh/bobrenjc93/533/head -> origin/gh/bobrenjc93/533/head 2025-07-17T08:34:50.5144986Z * [new branch] gh/bobrenjc93/533/orig -> origin/gh/bobrenjc93/533/orig 2025-07-17T08:34:50.5149488Z * [new branch] gh/bobrenjc93/534/base -> origin/gh/bobrenjc93/534/base 2025-07-17T08:34:50.5152815Z * [new branch] gh/bobrenjc93/534/head -> origin/gh/bobrenjc93/534/head 2025-07-17T08:34:50.5156365Z * [new branch] gh/bobrenjc93/534/orig -> origin/gh/bobrenjc93/534/orig 2025-07-17T08:34:50.5160851Z * [new branch] gh/bobrenjc93/535/base -> origin/gh/bobrenjc93/535/base 2025-07-17T08:34:50.5164116Z * [new branch] gh/bobrenjc93/535/head -> origin/gh/bobrenjc93/535/head 2025-07-17T08:34:50.5167466Z * [new branch] gh/bobrenjc93/535/orig -> origin/gh/bobrenjc93/535/orig 2025-07-17T08:34:50.5172062Z * [new branch] gh/bobrenjc93/536/base -> origin/gh/bobrenjc93/536/base 2025-07-17T08:34:50.5175403Z * [new branch] gh/bobrenjc93/536/head -> origin/gh/bobrenjc93/536/head 2025-07-17T08:34:50.5178843Z * [new branch] gh/bobrenjc93/536/orig -> origin/gh/bobrenjc93/536/orig 2025-07-17T08:34:50.5184344Z * [new branch] gh/briancoutinho/2/base -> origin/gh/briancoutinho/2/base 2025-07-17T08:34:50.5187782Z * [new branch] gh/briancoutinho/2/head -> origin/gh/briancoutinho/2/head 2025-07-17T08:34:50.5193341Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-07-17T08:34:50.5196984Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-07-17T08:34:50.5201805Z * [new branch] gh/c00w/38/base -> origin/gh/c00w/38/base 2025-07-17T08:34:50.5205083Z * [new branch] gh/c00w/38/head -> origin/gh/c00w/38/head 2025-07-17T08:34:50.5208465Z * [new branch] gh/c00w/38/orig -> origin/gh/c00w/38/orig 2025-07-17T08:34:50.5213036Z * [new branch] gh/c00w/48/base -> origin/gh/c00w/48/base 2025-07-17T08:34:50.5216362Z * [new branch] gh/c00w/48/head -> origin/gh/c00w/48/head 2025-07-17T08:34:50.5219830Z * [new branch] gh/c00w/48/orig -> origin/gh/c00w/48/orig 2025-07-17T08:34:50.5224433Z * [new branch] gh/c00w/49/base -> origin/gh/c00w/49/base 2025-07-17T08:34:50.5228140Z * [new branch] gh/c00w/49/head -> origin/gh/c00w/49/head 2025-07-17T08:34:50.5231317Z * [new branch] gh/c00w/49/orig -> origin/gh/c00w/49/orig 2025-07-17T08:34:50.5236196Z * [new branch] gh/c00w/50/base -> origin/gh/c00w/50/base 2025-07-17T08:34:50.5239612Z * [new branch] gh/c00w/50/head -> origin/gh/c00w/50/head 2025-07-17T08:34:50.5242969Z * [new branch] gh/c00w/50/orig -> origin/gh/c00w/50/orig 2025-07-17T08:34:50.5247498Z * [new branch] gh/c00w/51/base -> origin/gh/c00w/51/base 2025-07-17T08:34:50.5250868Z * [new branch] gh/c00w/51/head -> origin/gh/c00w/51/head 2025-07-17T08:34:50.5254228Z * [new branch] gh/c00w/51/orig -> origin/gh/c00w/51/orig 2025-07-17T08:34:50.5259662Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-07-17T08:34:50.5263156Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-07-17T08:34:50.5266503Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-07-17T08:34:50.5271043Z * [new branch] gh/clee2000/2/base -> origin/gh/clee2000/2/base 2025-07-17T08:34:50.5274436Z * [new branch] gh/clee2000/2/head -> origin/gh/clee2000/2/head 2025-07-17T08:34:50.5277899Z * [new branch] gh/clee2000/2/orig -> origin/gh/clee2000/2/orig 2025-07-17T08:34:50.5282341Z * [new branch] gh/clee2000/3/base -> origin/gh/clee2000/3/base 2025-07-17T08:34:50.5285666Z * [new branch] gh/clee2000/3/head -> origin/gh/clee2000/3/head 2025-07-17T08:34:50.5289040Z * [new branch] gh/clee2000/3/orig -> origin/gh/clee2000/3/orig 2025-07-17T08:34:50.5294597Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-07-17T08:34:50.5298061Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-07-17T08:34:50.5302767Z * [new branch] gh/coconutruben/10/base -> origin/gh/coconutruben/10/base 2025-07-17T08:34:50.5306428Z * [new branch] gh/coconutruben/10/head -> origin/gh/coconutruben/10/head 2025-07-17T08:34:50.5310027Z * [new branch] gh/coconutruben/10/orig -> origin/gh/coconutruben/10/orig 2025-07-17T08:34:50.5314881Z * [new branch] gh/coconutruben/11/base -> origin/gh/coconutruben/11/base 2025-07-17T08:34:50.5318419Z * [new branch] gh/coconutruben/11/head -> origin/gh/coconutruben/11/head 2025-07-17T08:34:50.5321824Z * [new branch] gh/coconutruben/11/orig -> origin/gh/coconutruben/11/orig 2025-07-17T08:34:50.5326836Z * [new branch] gh/coconutruben/12/base -> origin/gh/coconutruben/12/base 2025-07-17T08:34:50.5330631Z * [new branch] gh/coconutruben/12/head -> origin/gh/coconutruben/12/head 2025-07-17T08:34:50.5334252Z * [new branch] gh/coconutruben/12/orig -> origin/gh/coconutruben/12/orig 2025-07-17T08:34:50.5338845Z * [new branch] gh/coconutruben/13/base -> origin/gh/coconutruben/13/base 2025-07-17T08:34:50.5342228Z * [new branch] gh/coconutruben/13/head -> origin/gh/coconutruben/13/head 2025-07-17T08:34:50.5345714Z * [new branch] gh/coconutruben/13/orig -> origin/gh/coconutruben/13/orig 2025-07-17T08:34:50.5350424Z * [new branch] gh/coconutruben/14/base -> origin/gh/coconutruben/14/base 2025-07-17T08:34:50.5354019Z * [new branch] gh/coconutruben/14/head -> origin/gh/coconutruben/14/head 2025-07-17T08:34:50.5357515Z * [new branch] gh/coconutruben/14/orig -> origin/gh/coconutruben/14/orig 2025-07-17T08:34:50.5362347Z * [new branch] gh/coconutruben/15/base -> origin/gh/coconutruben/15/base 2025-07-17T08:34:50.5366206Z * [new branch] gh/coconutruben/15/head -> origin/gh/coconutruben/15/head 2025-07-17T08:34:50.5369450Z * [new branch] gh/coconutruben/15/orig -> origin/gh/coconutruben/15/orig 2025-07-17T08:34:50.5374109Z * [new branch] gh/coconutruben/16/base -> origin/gh/coconutruben/16/base 2025-07-17T08:34:50.5377405Z * [new branch] gh/coconutruben/16/head -> origin/gh/coconutruben/16/head 2025-07-17T08:34:50.5380759Z * [new branch] gh/coconutruben/16/orig -> origin/gh/coconutruben/16/orig 2025-07-17T08:34:50.5385578Z * [new branch] gh/coconutruben/17/base -> origin/gh/coconutruben/17/base 2025-07-17T08:34:50.5389107Z * [new branch] gh/coconutruben/17/head -> origin/gh/coconutruben/17/head 2025-07-17T08:34:50.5392603Z * [new branch] gh/coconutruben/17/orig -> origin/gh/coconutruben/17/orig 2025-07-17T08:34:50.5399990Z * [new branch] gh/coconutruben/18/base -> origin/gh/coconutruben/18/base 2025-07-17T08:34:50.5403464Z * [new branch] gh/coconutruben/18/head -> origin/gh/coconutruben/18/head 2025-07-17T08:34:50.5406838Z * [new branch] gh/coconutruben/18/orig -> origin/gh/coconutruben/18/orig 2025-07-17T08:34:50.5411517Z * [new branch] gh/coconutruben/19/base -> origin/gh/coconutruben/19/base 2025-07-17T08:34:50.5414950Z * [new branch] gh/coconutruben/19/head -> origin/gh/coconutruben/19/head 2025-07-17T08:34:50.5418398Z * [new branch] gh/coconutruben/19/orig -> origin/gh/coconutruben/19/orig 2025-07-17T08:34:50.5423166Z * [new branch] gh/coconutruben/2/base -> origin/gh/coconutruben/2/base 2025-07-17T08:34:50.5426587Z * [new branch] gh/coconutruben/2/head -> origin/gh/coconutruben/2/head 2025-07-17T08:34:50.5429990Z * [new branch] gh/coconutruben/2/orig -> origin/gh/coconutruben/2/orig 2025-07-17T08:34:50.5434926Z * [new branch] gh/coconutruben/20/base -> origin/gh/coconutruben/20/base 2025-07-17T08:34:50.5438505Z * [new branch] gh/coconutruben/20/head -> origin/gh/coconutruben/20/head 2025-07-17T08:34:50.5441806Z * [new branch] gh/coconutruben/20/orig -> origin/gh/coconutruben/20/orig 2025-07-17T08:34:50.5446610Z * [new branch] gh/coconutruben/21/base -> origin/gh/coconutruben/21/base 2025-07-17T08:34:50.5449869Z * [new branch] gh/coconutruben/21/head -> origin/gh/coconutruben/21/head 2025-07-17T08:34:50.5453235Z * [new branch] gh/coconutruben/21/orig -> origin/gh/coconutruben/21/orig 2025-07-17T08:34:50.5457772Z * [new branch] gh/coconutruben/22/base -> origin/gh/coconutruben/22/base 2025-07-17T08:34:50.5461126Z * [new branch] gh/coconutruben/22/head -> origin/gh/coconutruben/22/head 2025-07-17T08:34:50.5464630Z * [new branch] gh/coconutruben/22/orig -> origin/gh/coconutruben/22/orig 2025-07-17T08:34:50.5469412Z * [new branch] gh/coconutruben/23/base -> origin/gh/coconutruben/23/base 2025-07-17T08:34:50.5472831Z * [new branch] gh/coconutruben/23/head -> origin/gh/coconutruben/23/head 2025-07-17T08:34:50.5476558Z * [new branch] gh/coconutruben/23/orig -> origin/gh/coconutruben/23/orig 2025-07-17T08:34:50.5480953Z * [new branch] gh/coconutruben/24/base -> origin/gh/coconutruben/24/base 2025-07-17T08:34:50.5484339Z * [new branch] gh/coconutruben/24/head -> origin/gh/coconutruben/24/head 2025-07-17T08:34:50.5487774Z * [new branch] gh/coconutruben/24/orig -> origin/gh/coconutruben/24/orig 2025-07-17T08:34:50.5492690Z * [new branch] gh/coconutruben/3/base -> origin/gh/coconutruben/3/base 2025-07-17T08:34:50.5496342Z * [new branch] gh/coconutruben/3/head -> origin/gh/coconutruben/3/head 2025-07-17T08:34:50.5500189Z * [new branch] gh/coconutruben/3/orig -> origin/gh/coconutruben/3/orig 2025-07-17T08:34:50.5504581Z * [new branch] gh/coconutruben/4/base -> origin/gh/coconutruben/4/base 2025-07-17T08:34:50.5508017Z * [new branch] gh/coconutruben/4/head -> origin/gh/coconutruben/4/head 2025-07-17T08:34:50.5511410Z * [new branch] gh/coconutruben/4/orig -> origin/gh/coconutruben/4/orig 2025-07-17T08:34:50.5516146Z * [new branch] gh/coconutruben/5/base -> origin/gh/coconutruben/5/base 2025-07-17T08:34:50.5519393Z * [new branch] gh/coconutruben/5/head -> origin/gh/coconutruben/5/head 2025-07-17T08:34:50.5522778Z * [new branch] gh/coconutruben/5/orig -> origin/gh/coconutruben/5/orig 2025-07-17T08:34:50.5527651Z * [new branch] gh/coconutruben/6/base -> origin/gh/coconutruben/6/base 2025-07-17T08:34:50.5531188Z * [new branch] gh/coconutruben/6/head -> origin/gh/coconutruben/6/head 2025-07-17T08:34:50.5534887Z * [new branch] gh/coconutruben/6/orig -> origin/gh/coconutruben/6/orig 2025-07-17T08:34:50.5539698Z * [new branch] gh/coconutruben/7/base -> origin/gh/coconutruben/7/base 2025-07-17T08:34:50.5543189Z * [new branch] gh/coconutruben/7/head -> origin/gh/coconutruben/7/head 2025-07-17T08:34:50.5546669Z * [new branch] gh/coconutruben/7/orig -> origin/gh/coconutruben/7/orig 2025-07-17T08:34:50.5551186Z * [new branch] gh/coconutruben/8/base -> origin/gh/coconutruben/8/base 2025-07-17T08:34:50.5554745Z * [new branch] gh/coconutruben/8/head -> origin/gh/coconutruben/8/head 2025-07-17T08:34:50.5558212Z * [new branch] gh/coconutruben/8/orig -> origin/gh/coconutruben/8/orig 2025-07-17T08:34:50.5563147Z * [new branch] gh/coconutruben/9/base -> origin/gh/coconutruben/9/base 2025-07-17T08:34:50.5566653Z * [new branch] gh/coconutruben/9/head -> origin/gh/coconutruben/9/head 2025-07-17T08:34:50.5570162Z * [new branch] gh/coconutruben/9/orig -> origin/gh/coconutruben/9/orig 2025-07-17T08:34:50.5575769Z * [new branch] gh/codingwithsurya/1/base -> origin/gh/codingwithsurya/1/base 2025-07-17T08:34:50.5579055Z * [new branch] gh/codingwithsurya/1/head -> origin/gh/codingwithsurya/1/head 2025-07-17T08:34:50.5582459Z * [new branch] gh/codingwithsurya/1/orig -> origin/gh/codingwithsurya/1/orig 2025-07-17T08:34:50.5587160Z * [new branch] gh/codingwithsurya/10/base -> origin/gh/codingwithsurya/10/base 2025-07-17T08:34:50.5590564Z * [new branch] gh/codingwithsurya/10/head -> origin/gh/codingwithsurya/10/head 2025-07-17T08:34:50.5593982Z * [new branch] gh/codingwithsurya/10/orig -> origin/gh/codingwithsurya/10/orig 2025-07-17T08:34:50.5598761Z * [new branch] gh/codingwithsurya/11/base -> origin/gh/codingwithsurya/11/base 2025-07-17T08:34:50.5602001Z * [new branch] gh/codingwithsurya/11/head -> origin/gh/codingwithsurya/11/head 2025-07-17T08:34:50.5605232Z * [new branch] gh/codingwithsurya/11/orig -> origin/gh/codingwithsurya/11/orig 2025-07-17T08:34:50.5610004Z * [new branch] gh/codingwithsurya/12/base -> origin/gh/codingwithsurya/12/base 2025-07-17T08:34:50.5613355Z * [new branch] gh/codingwithsurya/12/head -> origin/gh/codingwithsurya/12/head 2025-07-17T08:34:50.5616707Z * [new branch] gh/codingwithsurya/12/orig -> origin/gh/codingwithsurya/12/orig 2025-07-17T08:34:50.5621287Z * [new branch] gh/codingwithsurya/2/base -> origin/gh/codingwithsurya/2/base 2025-07-17T08:34:50.5624637Z * [new branch] gh/codingwithsurya/2/head -> origin/gh/codingwithsurya/2/head 2025-07-17T08:34:50.5628118Z * [new branch] gh/codingwithsurya/2/orig -> origin/gh/codingwithsurya/2/orig 2025-07-17T08:34:50.5632823Z * [new branch] gh/codingwithsurya/3/base -> origin/gh/codingwithsurya/3/base 2025-07-17T08:34:50.5636305Z * [new branch] gh/codingwithsurya/3/head -> origin/gh/codingwithsurya/3/head 2025-07-17T08:34:50.5639583Z * [new branch] gh/codingwithsurya/3/orig -> origin/gh/codingwithsurya/3/orig 2025-07-17T08:34:50.5644042Z * [new branch] gh/codingwithsurya/4/base -> origin/gh/codingwithsurya/4/base 2025-07-17T08:34:50.5647532Z * [new branch] gh/codingwithsurya/4/head -> origin/gh/codingwithsurya/4/head 2025-07-17T08:34:50.5650895Z * [new branch] gh/codingwithsurya/4/orig -> origin/gh/codingwithsurya/4/orig 2025-07-17T08:34:50.5655588Z * [new branch] gh/codingwithsurya/5/base -> origin/gh/codingwithsurya/5/base 2025-07-17T08:34:50.5658907Z * [new branch] gh/codingwithsurya/5/head -> origin/gh/codingwithsurya/5/head 2025-07-17T08:34:50.5662257Z * [new branch] gh/codingwithsurya/5/orig -> origin/gh/codingwithsurya/5/orig 2025-07-17T08:34:50.5666902Z * [new branch] gh/codingwithsurya/6/base -> origin/gh/codingwithsurya/6/base 2025-07-17T08:34:50.5670383Z * [new branch] gh/codingwithsurya/6/head -> origin/gh/codingwithsurya/6/head 2025-07-17T08:34:50.5673594Z * [new branch] gh/codingwithsurya/6/orig -> origin/gh/codingwithsurya/6/orig 2025-07-17T08:34:50.5678721Z * [new branch] gh/codingwithsurya/7/base -> origin/gh/codingwithsurya/7/base 2025-07-17T08:34:50.5682031Z * [new branch] gh/codingwithsurya/7/head -> origin/gh/codingwithsurya/7/head 2025-07-17T08:34:50.5685558Z * [new branch] gh/codingwithsurya/7/orig -> origin/gh/codingwithsurya/7/orig 2025-07-17T08:34:50.5690203Z * [new branch] gh/codingwithsurya/8/base -> origin/gh/codingwithsurya/8/base 2025-07-17T08:34:50.5694693Z * [new branch] gh/codingwithsurya/8/head -> origin/gh/codingwithsurya/8/head 2025-07-17T08:34:50.5698029Z * [new branch] gh/codingwithsurya/8/orig -> origin/gh/codingwithsurya/8/orig 2025-07-17T08:34:50.5703323Z * [new branch] gh/codingwithsurya/9/base -> origin/gh/codingwithsurya/9/base 2025-07-17T08:34:50.5705944Z * [new branch] gh/codingwithsurya/9/head -> origin/gh/codingwithsurya/9/head 2025-07-17T08:34:50.5709206Z * [new branch] gh/codingwithsurya/9/orig -> origin/gh/codingwithsurya/9/orig 2025-07-17T08:34:50.5714774Z * [new branch] gh/davidberard98/347/base -> origin/gh/davidberard98/347/base 2025-07-17T08:34:50.5718476Z * [new branch] gh/davidberard98/347/head -> origin/gh/davidberard98/347/head 2025-07-17T08:34:50.5721823Z * [new branch] gh/davidberard98/347/orig -> origin/gh/davidberard98/347/orig 2025-07-17T08:34:50.5726405Z * [new branch] gh/davidberard98/351/base -> origin/gh/davidberard98/351/base 2025-07-17T08:34:50.5729745Z * [new branch] gh/davidberard98/351/head -> origin/gh/davidberard98/351/head 2025-07-17T08:34:50.5733099Z * [new branch] gh/davidberard98/351/orig -> origin/gh/davidberard98/351/orig 2025-07-17T08:34:50.5737521Z * [new branch] gh/davidberard98/353/base -> origin/gh/davidberard98/353/base 2025-07-17T08:34:50.5740936Z * [new branch] gh/davidberard98/353/head -> origin/gh/davidberard98/353/head 2025-07-17T08:34:50.5744185Z * [new branch] gh/davidberard98/353/orig -> origin/gh/davidberard98/353/orig 2025-07-17T08:34:50.5748709Z * [new branch] gh/davidberard98/356/base -> origin/gh/davidberard98/356/base 2025-07-17T08:34:50.5752057Z * [new branch] gh/davidberard98/356/head -> origin/gh/davidberard98/356/head 2025-07-17T08:34:50.5755685Z * [new branch] gh/davidberard98/356/orig -> origin/gh/davidberard98/356/orig 2025-07-17T08:34:50.5760297Z * [new branch] gh/davidberard98/370/base -> origin/gh/davidberard98/370/base 2025-07-17T08:34:50.5763912Z * [new branch] gh/davidberard98/370/head -> origin/gh/davidberard98/370/head 2025-07-17T08:34:50.5767189Z * [new branch] gh/davidberard98/370/orig -> origin/gh/davidberard98/370/orig 2025-07-17T08:34:50.5771954Z * [new branch] gh/davidberard98/375/base -> origin/gh/davidberard98/375/base 2025-07-17T08:34:50.5775330Z * [new branch] gh/davidberard98/375/head -> origin/gh/davidberard98/375/head 2025-07-17T08:34:50.5778692Z * [new branch] gh/davidberard98/375/orig -> origin/gh/davidberard98/375/orig 2025-07-17T08:34:50.5783286Z * [new branch] gh/davidberard98/378/base -> origin/gh/davidberard98/378/base 2025-07-17T08:34:50.5786667Z * [new branch] gh/davidberard98/378/head -> origin/gh/davidberard98/378/head 2025-07-17T08:34:50.5790017Z * [new branch] gh/davidberard98/378/orig -> origin/gh/davidberard98/378/orig 2025-07-17T08:34:50.5794800Z * [new branch] gh/davidberard98/379/base -> origin/gh/davidberard98/379/base 2025-07-17T08:34:50.5798279Z * [new branch] gh/davidberard98/379/head -> origin/gh/davidberard98/379/head 2025-07-17T08:34:50.5801525Z * [new branch] gh/davidberard98/379/orig -> origin/gh/davidberard98/379/orig 2025-07-17T08:34:50.5806224Z * [new branch] gh/davidberard98/380/base -> origin/gh/davidberard98/380/base 2025-07-17T08:34:50.5809641Z * [new branch] gh/davidberard98/380/head -> origin/gh/davidberard98/380/head 2025-07-17T08:34:50.5812931Z * [new branch] gh/davidberard98/380/orig -> origin/gh/davidberard98/380/orig 2025-07-17T08:34:50.5817402Z * [new branch] gh/davidberard98/381/base -> origin/gh/davidberard98/381/base 2025-07-17T08:34:50.5820722Z * [new branch] gh/davidberard98/381/head -> origin/gh/davidberard98/381/head 2025-07-17T08:34:50.5824057Z * [new branch] gh/davidberard98/381/orig -> origin/gh/davidberard98/381/orig 2025-07-17T08:34:50.5828736Z * [new branch] gh/davidberard98/382/base -> origin/gh/davidberard98/382/base 2025-07-17T08:34:50.5832157Z * [new branch] gh/davidberard98/382/head -> origin/gh/davidberard98/382/head 2025-07-17T08:34:50.5835680Z * [new branch] gh/davidberard98/382/orig -> origin/gh/davidberard98/382/orig 2025-07-17T08:34:50.5840371Z * [new branch] gh/davidberard98/383/base -> origin/gh/davidberard98/383/base 2025-07-17T08:34:50.5843827Z * [new branch] gh/davidberard98/383/head -> origin/gh/davidberard98/383/head 2025-07-17T08:34:50.5847191Z * [new branch] gh/davidberard98/383/orig -> origin/gh/davidberard98/383/orig 2025-07-17T08:34:50.5851641Z * [new branch] gh/davidberard98/384/base -> origin/gh/davidberard98/384/base 2025-07-17T08:34:50.5855030Z * [new branch] gh/davidberard98/384/head -> origin/gh/davidberard98/384/head 2025-07-17T08:34:50.5858350Z * [new branch] gh/davidberard98/384/orig -> origin/gh/davidberard98/384/orig 2025-07-17T08:34:50.5862967Z * [new branch] gh/davidberard98/385/base -> origin/gh/davidberard98/385/base 2025-07-17T08:34:50.5866311Z * [new branch] gh/davidberard98/385/head -> origin/gh/davidberard98/385/head 2025-07-17T08:34:50.5869681Z * [new branch] gh/davidberard98/385/orig -> origin/gh/davidberard98/385/orig 2025-07-17T08:34:50.5874667Z * [new branch] gh/davidberard98/386/base -> origin/gh/davidberard98/386/base 2025-07-17T08:34:50.5877917Z * [new branch] gh/davidberard98/386/head -> origin/gh/davidberard98/386/head 2025-07-17T08:34:50.5881186Z * [new branch] gh/davidberard98/386/orig -> origin/gh/davidberard98/386/orig 2025-07-17T08:34:50.5885879Z * [new branch] gh/davidberard98/387/base -> origin/gh/davidberard98/387/base 2025-07-17T08:34:50.5890677Z * [new branch] gh/davidberard98/387/head -> origin/gh/davidberard98/387/head 2025-07-17T08:34:50.5892733Z * [new branch] gh/davidberard98/387/orig -> origin/gh/davidberard98/387/orig 2025-07-17T08:34:50.5897105Z * [new branch] gh/davidberard98/388/base -> origin/gh/davidberard98/388/base 2025-07-17T08:34:50.5900675Z * [new branch] gh/davidberard98/388/head -> origin/gh/davidberard98/388/head 2025-07-17T08:34:50.5904027Z * [new branch] gh/davidberard98/388/orig -> origin/gh/davidberard98/388/orig 2025-07-17T08:34:50.5908644Z * [new branch] gh/davidberard98/389/base -> origin/gh/davidberard98/389/base 2025-07-17T08:34:50.5912179Z * [new branch] gh/davidberard98/389/head -> origin/gh/davidberard98/389/head 2025-07-17T08:34:50.5916381Z * [new branch] gh/davidberard98/389/orig -> origin/gh/davidberard98/389/orig 2025-07-17T08:34:50.5922182Z * [new branch] gh/davidberard98/390/base -> origin/gh/davidberard98/390/base 2025-07-17T08:34:50.5924760Z * [new branch] gh/davidberard98/390/head -> origin/gh/davidberard98/390/head 2025-07-17T08:34:50.5928164Z * [new branch] gh/davidberard98/390/orig -> origin/gh/davidberard98/390/orig 2025-07-17T08:34:50.5933650Z * [new branch] gh/desertfire/559/base -> origin/gh/desertfire/559/base 2025-07-17T08:34:50.5937025Z * [new branch] gh/desertfire/559/head -> origin/gh/desertfire/559/head 2025-07-17T08:34:50.5940323Z * [new branch] gh/desertfire/559/orig -> origin/gh/desertfire/559/orig 2025-07-17T08:34:50.5945020Z * [new branch] gh/desertfire/570/base -> origin/gh/desertfire/570/base 2025-07-17T08:34:50.5948482Z * [new branch] gh/desertfire/570/head -> origin/gh/desertfire/570/head 2025-07-17T08:34:50.5951810Z * [new branch] gh/desertfire/570/orig -> origin/gh/desertfire/570/orig 2025-07-17T08:34:50.5957895Z * [new branch] gh/desertfire/572/base -> origin/gh/desertfire/572/base 2025-07-17T08:34:50.5961188Z * [new branch] gh/desertfire/572/head -> origin/gh/desertfire/572/head 2025-07-17T08:34:50.5964596Z * [new branch] gh/desertfire/572/orig -> origin/gh/desertfire/572/orig 2025-07-17T08:34:50.5969227Z * [new branch] gh/desertfire/585/base -> origin/gh/desertfire/585/base 2025-07-17T08:34:50.5972557Z * [new branch] gh/desertfire/585/head -> origin/gh/desertfire/585/head 2025-07-17T08:34:50.5975901Z * [new branch] gh/desertfire/585/orig -> origin/gh/desertfire/585/orig 2025-07-17T08:34:50.5980576Z * [new branch] gh/desertfire/586/base -> origin/gh/desertfire/586/base 2025-07-17T08:34:50.5983979Z * [new branch] gh/desertfire/586/head -> origin/gh/desertfire/586/head 2025-07-17T08:34:50.5987358Z * [new branch] gh/desertfire/586/orig -> origin/gh/desertfire/586/orig 2025-07-17T08:34:50.5992067Z * [new branch] gh/desertfire/587/base -> origin/gh/desertfire/587/base 2025-07-17T08:34:50.5995667Z * [new branch] gh/desertfire/587/head -> origin/gh/desertfire/587/head 2025-07-17T08:34:50.5999017Z * [new branch] gh/desertfire/587/orig -> origin/gh/desertfire/587/orig 2025-07-17T08:34:50.6003554Z * [new branch] gh/desertfire/588/base -> origin/gh/desertfire/588/base 2025-07-17T08:34:50.6006936Z * [new branch] gh/desertfire/588/head -> origin/gh/desertfire/588/head 2025-07-17T08:34:50.6010303Z * [new branch] gh/desertfire/588/orig -> origin/gh/desertfire/588/orig 2025-07-17T08:34:50.6015777Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-07-17T08:34:50.6019150Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-07-17T08:34:50.6023714Z * [new branch] gh/dharakk/4/base -> origin/gh/dharakk/4/base 2025-07-17T08:34:50.6027238Z * [new branch] gh/dharakk/4/head -> origin/gh/dharakk/4/head 2025-07-17T08:34:50.6030363Z * [new branch] gh/dharakk/4/orig -> origin/gh/dharakk/4/orig 2025-07-17T08:34:50.6036125Z * [new branch] gh/drisspg/136/base -> origin/gh/drisspg/136/base 2025-07-17T08:34:50.6039428Z * [new branch] gh/drisspg/136/head -> origin/gh/drisspg/136/head 2025-07-17T08:34:50.6042750Z * [new branch] gh/drisspg/136/orig -> origin/gh/drisspg/136/orig 2025-07-17T08:34:50.6047238Z * [new branch] gh/drisspg/140/base -> origin/gh/drisspg/140/base 2025-07-17T08:34:50.6050605Z * [new branch] gh/drisspg/140/head -> origin/gh/drisspg/140/head 2025-07-17T08:34:50.6053921Z * [new branch] gh/drisspg/140/orig -> origin/gh/drisspg/140/orig 2025-07-17T08:34:50.6058440Z * [new branch] gh/drisspg/149/base -> origin/gh/drisspg/149/base 2025-07-17T08:34:50.6061734Z * [new branch] gh/drisspg/149/head -> origin/gh/drisspg/149/head 2025-07-17T08:34:50.6065095Z * [new branch] gh/drisspg/149/orig -> origin/gh/drisspg/149/orig 2025-07-17T08:34:50.6069637Z * [new branch] gh/drisspg/150/base -> origin/gh/drisspg/150/base 2025-07-17T08:34:50.6073039Z * [new branch] gh/drisspg/150/head -> origin/gh/drisspg/150/head 2025-07-17T08:34:50.6076593Z * [new branch] gh/drisspg/150/orig -> origin/gh/drisspg/150/orig 2025-07-17T08:34:50.6081162Z * [new branch] gh/drisspg/151/base -> origin/gh/drisspg/151/base 2025-07-17T08:34:50.6084445Z * [new branch] gh/drisspg/151/head -> origin/gh/drisspg/151/head 2025-07-17T08:34:50.6087758Z * [new branch] gh/drisspg/151/orig -> origin/gh/drisspg/151/orig 2025-07-17T08:34:50.6092322Z * [new branch] gh/drisspg/154/base -> origin/gh/drisspg/154/base 2025-07-17T08:34:50.6095684Z * [new branch] gh/drisspg/154/head -> origin/gh/drisspg/154/head 2025-07-17T08:34:50.6099122Z * [new branch] gh/drisspg/154/orig -> origin/gh/drisspg/154/orig 2025-07-17T08:34:50.6103635Z * [new branch] gh/drisspg/155/base -> origin/gh/drisspg/155/base 2025-07-17T08:34:50.6107141Z * [new branch] gh/drisspg/155/head -> origin/gh/drisspg/155/head 2025-07-17T08:34:50.6110427Z * [new branch] gh/drisspg/155/orig -> origin/gh/drisspg/155/orig 2025-07-17T08:34:50.6115207Z * [new branch] gh/drisspg/156/base -> origin/gh/drisspg/156/base 2025-07-17T08:34:50.6118574Z * [new branch] gh/drisspg/156/head -> origin/gh/drisspg/156/head 2025-07-17T08:34:50.6121756Z * [new branch] gh/drisspg/156/orig -> origin/gh/drisspg/156/orig 2025-07-17T08:34:50.6126431Z * [new branch] gh/drisspg/157/base -> origin/gh/drisspg/157/base 2025-07-17T08:34:50.6129788Z * [new branch] gh/drisspg/157/head -> origin/gh/drisspg/157/head 2025-07-17T08:34:50.6133121Z * [new branch] gh/drisspg/157/orig -> origin/gh/drisspg/157/orig 2025-07-17T08:34:50.6137528Z * [new branch] gh/drisspg/158/base -> origin/gh/drisspg/158/base 2025-07-17T08:34:50.6140872Z * [new branch] gh/drisspg/158/head -> origin/gh/drisspg/158/head 2025-07-17T08:34:50.6144164Z * [new branch] gh/drisspg/158/orig -> origin/gh/drisspg/158/orig 2025-07-17T08:34:50.6148652Z * [new branch] gh/drisspg/159/base -> origin/gh/drisspg/159/base 2025-07-17T08:34:50.6151995Z * [new branch] gh/drisspg/159/head -> origin/gh/drisspg/159/head 2025-07-17T08:34:50.6155927Z * [new branch] gh/drisspg/159/orig -> origin/gh/drisspg/159/orig 2025-07-17T08:34:50.6160728Z * [new branch] gh/drisspg/160/base -> origin/gh/drisspg/160/base 2025-07-17T08:34:50.6163840Z * [new branch] gh/drisspg/160/head -> origin/gh/drisspg/160/head 2025-07-17T08:34:50.6167190Z * [new branch] gh/drisspg/160/orig -> origin/gh/drisspg/160/orig 2025-07-17T08:34:50.6171782Z * [new branch] gh/drisspg/161/base -> origin/gh/drisspg/161/base 2025-07-17T08:34:50.6175054Z * [new branch] gh/drisspg/161/head -> origin/gh/drisspg/161/head 2025-07-17T08:34:50.6178377Z * [new branch] gh/drisspg/161/orig -> origin/gh/drisspg/161/orig 2025-07-17T08:34:50.6182939Z * [new branch] gh/drisspg/162/base -> origin/gh/drisspg/162/base 2025-07-17T08:34:50.6186258Z * [new branch] gh/drisspg/162/head -> origin/gh/drisspg/162/head 2025-07-17T08:34:50.6189636Z * [new branch] gh/drisspg/162/orig -> origin/gh/drisspg/162/orig 2025-07-17T08:34:50.6194231Z * [new branch] gh/drisspg/163/base -> origin/gh/drisspg/163/base 2025-07-17T08:34:50.6198842Z * [new branch] gh/drisspg/163/head -> origin/gh/drisspg/163/head 2025-07-17T08:34:50.6202231Z * [new branch] gh/drisspg/163/orig -> origin/gh/drisspg/163/orig 2025-07-17T08:34:50.6206892Z * [new branch] gh/drisspg/164/base -> origin/gh/drisspg/164/base 2025-07-17T08:34:50.6210328Z * [new branch] gh/drisspg/164/head -> origin/gh/drisspg/164/head 2025-07-17T08:34:50.6213578Z * [new branch] gh/drisspg/164/orig -> origin/gh/drisspg/164/orig 2025-07-17T08:34:50.6218060Z * [new branch] gh/drisspg/165/base -> origin/gh/drisspg/165/base 2025-07-17T08:34:50.6221400Z * [new branch] gh/drisspg/165/head -> origin/gh/drisspg/165/head 2025-07-17T08:34:50.6224645Z * [new branch] gh/drisspg/165/orig -> origin/gh/drisspg/165/orig 2025-07-17T08:34:50.6229308Z * [new branch] gh/drisspg/166/base -> origin/gh/drisspg/166/base 2025-07-17T08:34:50.6232613Z * [new branch] gh/drisspg/166/head -> origin/gh/drisspg/166/head 2025-07-17T08:34:50.6236188Z * [new branch] gh/drisspg/166/orig -> origin/gh/drisspg/166/orig 2025-07-17T08:34:50.6240690Z * [new branch] gh/drisspg/167/base -> origin/gh/drisspg/167/base 2025-07-17T08:34:50.6244029Z * [new branch] gh/drisspg/167/head -> origin/gh/drisspg/167/head 2025-07-17T08:34:50.6247442Z * [new branch] gh/drisspg/167/orig -> origin/gh/drisspg/167/orig 2025-07-17T08:34:50.6251979Z * [new branch] gh/drisspg/168/base -> origin/gh/drisspg/168/base 2025-07-17T08:34:50.6255288Z * [new branch] gh/drisspg/168/head -> origin/gh/drisspg/168/head 2025-07-17T08:34:50.6258608Z * [new branch] gh/drisspg/168/orig -> origin/gh/drisspg/168/orig 2025-07-17T08:34:50.6263183Z * [new branch] gh/drisspg/169/base -> origin/gh/drisspg/169/base 2025-07-17T08:34:50.6266470Z * [new branch] gh/drisspg/169/head -> origin/gh/drisspg/169/head 2025-07-17T08:34:50.6269779Z * [new branch] gh/drisspg/169/orig -> origin/gh/drisspg/169/orig 2025-07-17T08:34:50.6274411Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-07-17T08:34:50.6277863Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-07-17T08:34:50.6281108Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-07-17T08:34:50.6285635Z * [new branch] gh/drisspg/171/base -> origin/gh/drisspg/171/base 2025-07-17T08:34:50.6288933Z * [new branch] gh/drisspg/171/head -> origin/gh/drisspg/171/head 2025-07-17T08:34:50.6292395Z * [new branch] gh/drisspg/171/orig -> origin/gh/drisspg/171/orig 2025-07-17T08:34:50.6298319Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-07-17T08:34:50.6301430Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-07-17T08:34:50.6307098Z * [new branch] gh/eellison/779/base -> origin/gh/eellison/779/base 2025-07-17T08:34:50.6310367Z * [new branch] gh/eellison/779/head -> origin/gh/eellison/779/head 2025-07-17T08:34:50.6313719Z * [new branch] gh/eellison/779/orig -> origin/gh/eellison/779/orig 2025-07-17T08:34:50.6318447Z * [new branch] gh/eellison/781/base -> origin/gh/eellison/781/base 2025-07-17T08:34:50.6321745Z * [new branch] gh/eellison/781/head -> origin/gh/eellison/781/head 2025-07-17T08:34:50.6325062Z * [new branch] gh/eellison/781/orig -> origin/gh/eellison/781/orig 2025-07-17T08:34:50.6329585Z * [new branch] gh/eellison/784/base -> origin/gh/eellison/784/base 2025-07-17T08:34:50.6333031Z * [new branch] gh/eellison/784/head -> origin/gh/eellison/784/head 2025-07-17T08:34:50.6336422Z * [new branch] gh/eellison/784/orig -> origin/gh/eellison/784/orig 2025-07-17T08:34:50.6341146Z * [new branch] gh/eellison/785/base -> origin/gh/eellison/785/base 2025-07-17T08:34:50.6344513Z * [new branch] gh/eellison/785/head -> origin/gh/eellison/785/head 2025-07-17T08:34:50.6347864Z * [new branch] gh/eellison/785/orig -> origin/gh/eellison/785/orig 2025-07-17T08:34:50.6352367Z * [new branch] gh/eellison/789/base -> origin/gh/eellison/789/base 2025-07-17T08:34:50.6355967Z * [new branch] gh/eellison/789/head -> origin/gh/eellison/789/head 2025-07-17T08:34:50.6359259Z * [new branch] gh/eellison/789/orig -> origin/gh/eellison/789/orig 2025-07-17T08:34:50.6363730Z * [new branch] gh/eellison/800/base -> origin/gh/eellison/800/base 2025-07-17T08:34:50.6367032Z * [new branch] gh/eellison/800/head -> origin/gh/eellison/800/head 2025-07-17T08:34:50.6370364Z * [new branch] gh/eellison/800/orig -> origin/gh/eellison/800/orig 2025-07-17T08:34:50.6374850Z * [new branch] gh/eellison/801/base -> origin/gh/eellison/801/base 2025-07-17T08:34:50.6378283Z * [new branch] gh/eellison/801/head -> origin/gh/eellison/801/head 2025-07-17T08:34:50.6381588Z * [new branch] gh/eellison/801/orig -> origin/gh/eellison/801/orig 2025-07-17T08:34:50.6386282Z * [new branch] gh/eellison/802/base -> origin/gh/eellison/802/base 2025-07-17T08:34:50.6389608Z * [new branch] gh/eellison/802/head -> origin/gh/eellison/802/head 2025-07-17T08:34:50.6392906Z * [new branch] gh/eellison/802/orig -> origin/gh/eellison/802/orig 2025-07-17T08:34:50.6397845Z * [new branch] gh/eellison/805/base -> origin/gh/eellison/805/base 2025-07-17T08:34:50.6401084Z * [new branch] gh/eellison/805/head -> origin/gh/eellison/805/head 2025-07-17T08:34:50.6404458Z * [new branch] gh/eellison/805/orig -> origin/gh/eellison/805/orig 2025-07-17T08:34:50.6408996Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-07-17T08:34:50.6412299Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-07-17T08:34:50.6416098Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-07-17T08:34:50.6421270Z * [new branch] gh/etaf/117/base -> origin/gh/etaf/117/base 2025-07-17T08:34:50.6424686Z * [new branch] gh/etaf/117/head -> origin/gh/etaf/117/head 2025-07-17T08:34:50.6428109Z * [new branch] gh/etaf/117/orig -> origin/gh/etaf/117/orig 2025-07-17T08:34:50.6433276Z * [new branch] gh/etaf/130/base -> origin/gh/etaf/130/base 2025-07-17T08:34:50.6436369Z * [new branch] gh/etaf/130/head -> origin/gh/etaf/130/head 2025-07-17T08:34:50.6439648Z * [new branch] gh/etaf/130/orig -> origin/gh/etaf/130/orig 2025-07-17T08:34:50.6444154Z * [new branch] gh/etaf/132/base -> origin/gh/etaf/132/base 2025-07-17T08:34:50.6447551Z * [new branch] gh/etaf/132/head -> origin/gh/etaf/132/head 2025-07-17T08:34:50.6450869Z * [new branch] gh/etaf/132/orig -> origin/gh/etaf/132/orig 2025-07-17T08:34:50.6455348Z * [new branch] gh/etaf/134/base -> origin/gh/etaf/134/base 2025-07-17T08:34:50.6458761Z * [new branch] gh/etaf/134/head -> origin/gh/etaf/134/head 2025-07-17T08:34:50.6462109Z * [new branch] gh/etaf/134/orig -> origin/gh/etaf/134/orig 2025-07-17T08:34:50.6466811Z * [new branch] gh/etaf/137/base -> origin/gh/etaf/137/base 2025-07-17T08:34:50.6470166Z * [new branch] gh/etaf/137/head -> origin/gh/etaf/137/head 2025-07-17T08:34:50.6473511Z * [new branch] gh/etaf/137/orig -> origin/gh/etaf/137/orig 2025-07-17T08:34:50.6478327Z * [new branch] gh/etaf/138/base -> origin/gh/etaf/138/base 2025-07-17T08:34:50.6481586Z * [new branch] gh/etaf/138/head -> origin/gh/etaf/138/head 2025-07-17T08:34:50.6484903Z * [new branch] gh/etaf/138/orig -> origin/gh/etaf/138/orig 2025-07-17T08:34:50.6489739Z * [new branch] gh/etaf/139/base -> origin/gh/etaf/139/base 2025-07-17T08:34:50.6493191Z * [new branch] gh/etaf/139/head -> origin/gh/etaf/139/head 2025-07-17T08:34:50.6496936Z * [new branch] gh/etaf/139/orig -> origin/gh/etaf/139/orig 2025-07-17T08:34:50.6501412Z * [new branch] gh/etaf/140/base -> origin/gh/etaf/140/base 2025-07-17T08:34:50.6504697Z * [new branch] gh/etaf/140/head -> origin/gh/etaf/140/head 2025-07-17T08:34:50.6508192Z * [new branch] gh/etaf/140/orig -> origin/gh/etaf/140/orig 2025-07-17T08:34:50.6512575Z * [new branch] gh/etaf/141/base -> origin/gh/etaf/141/base 2025-07-17T08:34:50.6516251Z * [new branch] gh/etaf/141/head -> origin/gh/etaf/141/head 2025-07-17T08:34:50.6519604Z * [new branch] gh/etaf/141/orig -> origin/gh/etaf/141/orig 2025-07-17T08:34:50.6524090Z * [new branch] gh/etaf/142/base -> origin/gh/etaf/142/base 2025-07-17T08:34:50.6527385Z * [new branch] gh/etaf/142/head -> origin/gh/etaf/142/head 2025-07-17T08:34:50.6531882Z * [new branch] gh/etaf/143/base -> origin/gh/etaf/143/base 2025-07-17T08:34:50.6535310Z * [new branch] gh/etaf/143/head -> origin/gh/etaf/143/head 2025-07-17T08:34:50.6538640Z * [new branch] gh/etaf/143/orig -> origin/gh/etaf/143/orig 2025-07-17T08:34:50.6543193Z * [new branch] gh/etaf/144/base -> origin/gh/etaf/144/base 2025-07-17T08:34:50.6546531Z * [new branch] gh/etaf/144/head -> origin/gh/etaf/144/head 2025-07-17T08:34:50.6549860Z * [new branch] gh/etaf/144/orig -> origin/gh/etaf/144/orig 2025-07-17T08:34:50.6554571Z * [new branch] gh/etaf/145/base -> origin/gh/etaf/145/base 2025-07-17T08:34:50.6558099Z * [new branch] gh/etaf/145/head -> origin/gh/etaf/145/head 2025-07-17T08:34:50.6562476Z * [new branch] gh/etaf/146/base -> origin/gh/etaf/146/base 2025-07-17T08:34:50.6565803Z * [new branch] gh/etaf/146/head -> origin/gh/etaf/146/head 2025-07-17T08:34:50.6571660Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-07-17T08:34:50.6574850Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-07-17T08:34:50.6578183Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-07-17T08:34:50.6582747Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-07-17T08:34:50.6586090Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-07-17T08:34:50.6589466Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-07-17T08:34:50.6594038Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-07-17T08:34:50.6597527Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-07-17T08:34:50.6600914Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-07-17T08:34:50.6605510Z * [new branch] gh/ezyang/3068/base -> origin/gh/ezyang/3068/base 2025-07-17T08:34:50.6608900Z * [new branch] gh/ezyang/3068/head -> origin/gh/ezyang/3068/head 2025-07-17T08:34:50.6612230Z * [new branch] gh/ezyang/3068/orig -> origin/gh/ezyang/3068/orig 2025-07-17T08:34:50.6616901Z * [new branch] gh/ezyang/3071/base -> origin/gh/ezyang/3071/base 2025-07-17T08:34:50.6620208Z * [new branch] gh/ezyang/3071/head -> origin/gh/ezyang/3071/head 2025-07-17T08:34:50.6623545Z * [new branch] gh/ezyang/3071/orig -> origin/gh/ezyang/3071/orig 2025-07-17T08:34:50.6628053Z * [new branch] gh/ezyang/3072/base -> origin/gh/ezyang/3072/base 2025-07-17T08:34:50.6631342Z * [new branch] gh/ezyang/3072/head -> origin/gh/ezyang/3072/head 2025-07-17T08:34:50.6634764Z * [new branch] gh/ezyang/3072/orig -> origin/gh/ezyang/3072/orig 2025-07-17T08:34:50.6639391Z * [new branch] gh/ezyang/3074/base -> origin/gh/ezyang/3074/base 2025-07-17T08:34:50.6642775Z * [new branch] gh/ezyang/3074/head -> origin/gh/ezyang/3074/head 2025-07-17T08:34:50.6646132Z * [new branch] gh/ezyang/3074/orig -> origin/gh/ezyang/3074/orig 2025-07-17T08:34:50.6650791Z * [new branch] gh/ezyang/3078/base -> origin/gh/ezyang/3078/base 2025-07-17T08:34:50.6654161Z * [new branch] gh/ezyang/3078/head -> origin/gh/ezyang/3078/head 2025-07-17T08:34:50.6657485Z * [new branch] gh/ezyang/3078/orig -> origin/gh/ezyang/3078/orig 2025-07-17T08:34:50.6661967Z * [new branch] gh/ezyang/3079/base -> origin/gh/ezyang/3079/base 2025-07-17T08:34:50.6665608Z * [new branch] gh/ezyang/3079/head -> origin/gh/ezyang/3079/head 2025-07-17T08:34:50.6668677Z * [new branch] gh/ezyang/3079/orig -> origin/gh/ezyang/3079/orig 2025-07-17T08:34:50.6673206Z * [new branch] gh/ezyang/3080/base -> origin/gh/ezyang/3080/base 2025-07-17T08:34:50.6676833Z * [new branch] gh/ezyang/3080/head -> origin/gh/ezyang/3080/head 2025-07-17T08:34:50.6681309Z * [new branch] gh/ezyang/3080/orig -> origin/gh/ezyang/3080/orig 2025-07-17T08:34:50.6684924Z * [new branch] gh/ezyang/3081/base -> origin/gh/ezyang/3081/base 2025-07-17T08:34:50.6688527Z * [new branch] gh/ezyang/3081/head -> origin/gh/ezyang/3081/head 2025-07-17T08:34:50.6691665Z * [new branch] gh/ezyang/3081/orig -> origin/gh/ezyang/3081/orig 2025-07-17T08:34:50.6696488Z * [new branch] gh/ezyang/3082/base -> origin/gh/ezyang/3082/base 2025-07-17T08:34:50.6699578Z * [new branch] gh/ezyang/3082/head -> origin/gh/ezyang/3082/head 2025-07-17T08:34:50.6703101Z * [new branch] gh/ezyang/3082/orig -> origin/gh/ezyang/3082/orig 2025-07-17T08:34:50.6707471Z * [new branch] gh/ezyang/3083/base -> origin/gh/ezyang/3083/base 2025-07-17T08:34:50.6710791Z * [new branch] gh/ezyang/3083/head -> origin/gh/ezyang/3083/head 2025-07-17T08:34:50.6714174Z * [new branch] gh/ezyang/3083/orig -> origin/gh/ezyang/3083/orig 2025-07-17T08:34:50.6718892Z * [new branch] gh/ezyang/3084/base -> origin/gh/ezyang/3084/base 2025-07-17T08:34:50.6722181Z * [new branch] gh/ezyang/3084/head -> origin/gh/ezyang/3084/head 2025-07-17T08:34:50.6725555Z * [new branch] gh/ezyang/3084/orig -> origin/gh/ezyang/3084/orig 2025-07-17T08:34:50.6730034Z * [new branch] gh/ezyang/3085/base -> origin/gh/ezyang/3085/base 2025-07-17T08:34:50.6733372Z * [new branch] gh/ezyang/3085/head -> origin/gh/ezyang/3085/head 2025-07-17T08:34:50.6736758Z * [new branch] gh/ezyang/3085/orig -> origin/gh/ezyang/3085/orig 2025-07-17T08:34:50.6741350Z * [new branch] gh/ezyang/3086/base -> origin/gh/ezyang/3086/base 2025-07-17T08:34:50.6744695Z * [new branch] gh/ezyang/3086/head -> origin/gh/ezyang/3086/head 2025-07-17T08:34:50.6748002Z * [new branch] gh/ezyang/3086/orig -> origin/gh/ezyang/3086/orig 2025-07-17T08:34:50.6752557Z * [new branch] gh/ezyang/3087/base -> origin/gh/ezyang/3087/base 2025-07-17T08:34:50.6756726Z * [new branch] gh/ezyang/3087/head -> origin/gh/ezyang/3087/head 2025-07-17T08:34:50.6760043Z * [new branch] gh/ezyang/3087/orig -> origin/gh/ezyang/3087/orig 2025-07-17T08:34:50.6764617Z * [new branch] gh/ezyang/3088/base -> origin/gh/ezyang/3088/base 2025-07-17T08:34:50.6767952Z * [new branch] gh/ezyang/3088/head -> origin/gh/ezyang/3088/head 2025-07-17T08:34:50.6771353Z * [new branch] gh/ezyang/3088/orig -> origin/gh/ezyang/3088/orig 2025-07-17T08:34:50.6775899Z * [new branch] gh/ezyang/3089/base -> origin/gh/ezyang/3089/base 2025-07-17T08:34:50.6779208Z * [new branch] gh/ezyang/3089/head -> origin/gh/ezyang/3089/head 2025-07-17T08:34:50.6782670Z * [new branch] gh/ezyang/3089/orig -> origin/gh/ezyang/3089/orig 2025-07-17T08:34:50.6787341Z * [new branch] gh/ezyang/3090/base -> origin/gh/ezyang/3090/base 2025-07-17T08:34:50.6790642Z * [new branch] gh/ezyang/3090/head -> origin/gh/ezyang/3090/head 2025-07-17T08:34:50.6794063Z * [new branch] gh/ezyang/3090/orig -> origin/gh/ezyang/3090/orig 2025-07-17T08:34:50.6798720Z * [new branch] gh/ezyang/3091/base -> origin/gh/ezyang/3091/base 2025-07-17T08:34:50.6801994Z * [new branch] gh/ezyang/3091/head -> origin/gh/ezyang/3091/head 2025-07-17T08:34:50.6805409Z * [new branch] gh/ezyang/3091/orig -> origin/gh/ezyang/3091/orig 2025-07-17T08:34:50.6809944Z * [new branch] gh/ezyang/3092/base -> origin/gh/ezyang/3092/base 2025-07-17T08:34:50.6813343Z * [new branch] gh/ezyang/3092/head -> origin/gh/ezyang/3092/head 2025-07-17T08:34:50.6816703Z * [new branch] gh/ezyang/3092/orig -> origin/gh/ezyang/3092/orig 2025-07-17T08:34:50.6821291Z * [new branch] gh/ezyang/3093/base -> origin/gh/ezyang/3093/base 2025-07-17T08:34:50.6824636Z * [new branch] gh/ezyang/3093/head -> origin/gh/ezyang/3093/head 2025-07-17T08:34:50.6828034Z * [new branch] gh/ezyang/3093/orig -> origin/gh/ezyang/3093/orig 2025-07-17T08:34:50.6832697Z * [new branch] gh/ezyang/3094/base -> origin/gh/ezyang/3094/base 2025-07-17T08:34:50.6836216Z * [new branch] gh/ezyang/3094/head -> origin/gh/ezyang/3094/head 2025-07-17T08:34:50.6839772Z * [new branch] gh/ezyang/3094/orig -> origin/gh/ezyang/3094/orig 2025-07-17T08:34:50.6843937Z * [new branch] gh/ezyang/3095/base -> origin/gh/ezyang/3095/base 2025-07-17T08:34:50.6847256Z * [new branch] gh/ezyang/3095/head -> origin/gh/ezyang/3095/head 2025-07-17T08:34:50.6850558Z * [new branch] gh/ezyang/3095/orig -> origin/gh/ezyang/3095/orig 2025-07-17T08:34:50.6855076Z * [new branch] gh/ezyang/3096/base -> origin/gh/ezyang/3096/base 2025-07-17T08:34:50.6858364Z * [new branch] gh/ezyang/3096/head -> origin/gh/ezyang/3096/head 2025-07-17T08:34:50.6861662Z * [new branch] gh/ezyang/3096/orig -> origin/gh/ezyang/3096/orig 2025-07-17T08:34:50.6866164Z * [new branch] gh/ezyang/3097/base -> origin/gh/ezyang/3097/base 2025-07-17T08:34:50.6869715Z * [new branch] gh/ezyang/3097/head -> origin/gh/ezyang/3097/head 2025-07-17T08:34:50.6873228Z * [new branch] gh/ezyang/3097/orig -> origin/gh/ezyang/3097/orig 2025-07-17T08:34:50.6878044Z * [new branch] gh/ezyang/3098/base -> origin/gh/ezyang/3098/base 2025-07-17T08:34:50.6881314Z * [new branch] gh/ezyang/3098/head -> origin/gh/ezyang/3098/head 2025-07-17T08:34:50.6884626Z * [new branch] gh/ezyang/3098/orig -> origin/gh/ezyang/3098/orig 2025-07-17T08:34:50.6889116Z * [new branch] gh/ezyang/3099/base -> origin/gh/ezyang/3099/base 2025-07-17T08:34:50.6892435Z * [new branch] gh/ezyang/3099/head -> origin/gh/ezyang/3099/head 2025-07-17T08:34:50.6895902Z * [new branch] gh/ezyang/3099/orig -> origin/gh/ezyang/3099/orig 2025-07-17T08:34:50.6900432Z * [new branch] gh/ezyang/3100/base -> origin/gh/ezyang/3100/base 2025-07-17T08:34:50.6903842Z * [new branch] gh/ezyang/3100/head -> origin/gh/ezyang/3100/head 2025-07-17T08:34:50.6907159Z * [new branch] gh/ezyang/3100/orig -> origin/gh/ezyang/3100/orig 2025-07-17T08:34:50.6911723Z * [new branch] gh/ezyang/3101/base -> origin/gh/ezyang/3101/base 2025-07-17T08:34:50.6915205Z * [new branch] gh/ezyang/3101/head -> origin/gh/ezyang/3101/head 2025-07-17T08:34:50.6918788Z * [new branch] gh/ezyang/3101/orig -> origin/gh/ezyang/3101/orig 2025-07-17T08:34:50.6923391Z * [new branch] gh/ezyang/3102/base -> origin/gh/ezyang/3102/base 2025-07-17T08:34:50.6926672Z * [new branch] gh/ezyang/3102/head -> origin/gh/ezyang/3102/head 2025-07-17T08:34:50.6929998Z * [new branch] gh/ezyang/3102/orig -> origin/gh/ezyang/3102/orig 2025-07-17T08:34:50.6934587Z * [new branch] gh/ezyang/3103/base -> origin/gh/ezyang/3103/base 2025-07-17T08:34:50.6937916Z * [new branch] gh/ezyang/3103/head -> origin/gh/ezyang/3103/head 2025-07-17T08:34:50.6941263Z * [new branch] gh/ezyang/3103/orig -> origin/gh/ezyang/3103/orig 2025-07-17T08:34:50.6945785Z * [new branch] gh/ezyang/3104/base -> origin/gh/ezyang/3104/base 2025-07-17T08:34:50.6949098Z * [new branch] gh/ezyang/3104/head -> origin/gh/ezyang/3104/head 2025-07-17T08:34:50.6952502Z * [new branch] gh/ezyang/3104/orig -> origin/gh/ezyang/3104/orig 2025-07-17T08:34:50.6957224Z * [new branch] gh/ezyang/3105/base -> origin/gh/ezyang/3105/base 2025-07-17T08:34:50.6960478Z * [new branch] gh/ezyang/3105/head -> origin/gh/ezyang/3105/head 2025-07-17T08:34:50.6963936Z * [new branch] gh/ezyang/3105/orig -> origin/gh/ezyang/3105/orig 2025-07-17T08:34:50.6968475Z * [new branch] gh/ezyang/3106/base -> origin/gh/ezyang/3106/base 2025-07-17T08:34:50.6972139Z * [new branch] gh/ezyang/3106/head -> origin/gh/ezyang/3106/head 2025-07-17T08:34:50.6975263Z * [new branch] gh/ezyang/3106/orig -> origin/gh/ezyang/3106/orig 2025-07-17T08:34:50.6979759Z * [new branch] gh/ezyang/3107/base -> origin/gh/ezyang/3107/base 2025-07-17T08:34:50.6983080Z * [new branch] gh/ezyang/3107/head -> origin/gh/ezyang/3107/head 2025-07-17T08:34:50.6986397Z * [new branch] gh/ezyang/3107/orig -> origin/gh/ezyang/3107/orig 2025-07-17T08:34:50.6990926Z * [new branch] gh/ezyang/3108/base -> origin/gh/ezyang/3108/base 2025-07-17T08:34:50.6994372Z * [new branch] gh/ezyang/3108/head -> origin/gh/ezyang/3108/head 2025-07-17T08:34:50.6997869Z * [new branch] gh/ezyang/3108/orig -> origin/gh/ezyang/3108/orig 2025-07-17T08:34:50.7003354Z * [new branch] gh/fadara01/1/base -> origin/gh/fadara01/1/base 2025-07-17T08:34:50.7006808Z * [new branch] gh/fadara01/1/head -> origin/gh/fadara01/1/head 2025-07-17T08:34:50.7010156Z * [new branch] gh/fadara01/1/orig -> origin/gh/fadara01/1/orig 2025-07-17T08:34:50.7015646Z * [new branch] gh/fduwjj/146/base -> origin/gh/fduwjj/146/base 2025-07-17T08:34:50.7018925Z * [new branch] gh/fduwjj/146/head -> origin/gh/fduwjj/146/head 2025-07-17T08:34:50.7022309Z * [new branch] gh/fduwjj/146/orig -> origin/gh/fduwjj/146/orig 2025-07-17T08:34:50.7026935Z * [new branch] gh/fduwjj/150/base -> origin/gh/fduwjj/150/base 2025-07-17T08:34:50.7030284Z * [new branch] gh/fduwjj/150/head -> origin/gh/fduwjj/150/head 2025-07-17T08:34:50.7033638Z * [new branch] gh/fduwjj/150/orig -> origin/gh/fduwjj/150/orig 2025-07-17T08:34:50.7038749Z * [new branch] gh/fduwjj/151/base -> origin/gh/fduwjj/151/base 2025-07-17T08:34:50.7042036Z * [new branch] gh/fduwjj/151/head -> origin/gh/fduwjj/151/head 2025-07-17T08:34:50.7045480Z * [new branch] gh/fduwjj/151/orig -> origin/gh/fduwjj/151/orig 2025-07-17T08:34:50.7050169Z * [new branch] gh/fduwjj/152/base -> origin/gh/fduwjj/152/base 2025-07-17T08:34:50.7053526Z * [new branch] gh/fduwjj/152/head -> origin/gh/fduwjj/152/head 2025-07-17T08:34:50.7056878Z * [new branch] gh/fduwjj/152/orig -> origin/gh/fduwjj/152/orig 2025-07-17T08:34:50.7061684Z * [new branch] gh/fduwjj/153/base -> origin/gh/fduwjj/153/base 2025-07-17T08:34:50.7065150Z * [new branch] gh/fduwjj/153/head -> origin/gh/fduwjj/153/head 2025-07-17T08:34:50.7068519Z * [new branch] gh/fduwjj/153/orig -> origin/gh/fduwjj/153/orig 2025-07-17T08:34:50.7073243Z * [new branch] gh/fduwjj/154/base -> origin/gh/fduwjj/154/base 2025-07-17T08:34:50.7076864Z * [new branch] gh/fduwjj/154/head -> origin/gh/fduwjj/154/head 2025-07-17T08:34:50.7080208Z * [new branch] gh/fduwjj/154/orig -> origin/gh/fduwjj/154/orig 2025-07-17T08:34:50.7084716Z * [new branch] gh/fduwjj/155/base -> origin/gh/fduwjj/155/base 2025-07-17T08:34:50.7088156Z * [new branch] gh/fduwjj/155/head -> origin/gh/fduwjj/155/head 2025-07-17T08:34:50.7091472Z * [new branch] gh/fduwjj/155/orig -> origin/gh/fduwjj/155/orig 2025-07-17T08:34:50.7096246Z * [new branch] gh/fduwjj/156/base -> origin/gh/fduwjj/156/base 2025-07-17T08:34:50.7099586Z * [new branch] gh/fduwjj/156/head -> origin/gh/fduwjj/156/head 2025-07-17T08:34:50.7102931Z * [new branch] gh/fduwjj/156/orig -> origin/gh/fduwjj/156/orig 2025-07-17T08:34:50.7107521Z * [new branch] gh/fduwjj/157/base -> origin/gh/fduwjj/157/base 2025-07-17T08:34:50.7110631Z * [new branch] gh/fduwjj/157/head -> origin/gh/fduwjj/157/head 2025-07-17T08:34:50.7114032Z * [new branch] gh/fduwjj/157/orig -> origin/gh/fduwjj/157/orig 2025-07-17T08:34:50.7122163Z * [new branch] gh/fduwjj/158/base -> origin/gh/fduwjj/158/base 2025-07-17T08:34:50.7125631Z * [new branch] gh/fduwjj/158/head -> origin/gh/fduwjj/158/head 2025-07-17T08:34:50.7128951Z * [new branch] gh/fduwjj/158/orig -> origin/gh/fduwjj/158/orig 2025-07-17T08:34:50.7133427Z * [new branch] gh/fduwjj/159/base -> origin/gh/fduwjj/159/base 2025-07-17T08:34:50.7136896Z * [new branch] gh/fduwjj/159/head -> origin/gh/fduwjj/159/head 2025-07-17T08:34:50.7140224Z * [new branch] gh/fduwjj/159/orig -> origin/gh/fduwjj/159/orig 2025-07-17T08:34:50.7144639Z * [new branch] gh/fduwjj/160/base -> origin/gh/fduwjj/160/base 2025-07-17T08:34:50.7147993Z * [new branch] gh/fduwjj/160/head -> origin/gh/fduwjj/160/head 2025-07-17T08:34:50.7151284Z * [new branch] gh/fduwjj/160/orig -> origin/gh/fduwjj/160/orig 2025-07-17T08:34:50.7155979Z * [new branch] gh/fduwjj/161/base -> origin/gh/fduwjj/161/base 2025-07-17T08:34:50.7159393Z * [new branch] gh/fduwjj/161/head -> origin/gh/fduwjj/161/head 2025-07-17T08:34:50.7162646Z * [new branch] gh/fduwjj/161/orig -> origin/gh/fduwjj/161/orig 2025-07-17T08:34:50.7167187Z * [new branch] gh/fduwjj/162/base -> origin/gh/fduwjj/162/base 2025-07-17T08:34:50.7170519Z * [new branch] gh/fduwjj/162/head -> origin/gh/fduwjj/162/head 2025-07-17T08:34:50.7173888Z * [new branch] gh/fduwjj/162/orig -> origin/gh/fduwjj/162/orig 2025-07-17T08:34:50.7178633Z * [new branch] gh/fduwjj/163/base -> origin/gh/fduwjj/163/base 2025-07-17T08:34:50.7182038Z * [new branch] gh/fduwjj/163/head -> origin/gh/fduwjj/163/head 2025-07-17T08:34:50.7185403Z * [new branch] gh/fduwjj/163/orig -> origin/gh/fduwjj/163/orig 2025-07-17T08:34:50.7189911Z * [new branch] gh/fduwjj/164/base -> origin/gh/fduwjj/164/base 2025-07-17T08:34:50.7193369Z * [new branch] gh/fduwjj/164/head -> origin/gh/fduwjj/164/head 2025-07-17T08:34:50.7196934Z * [new branch] gh/fduwjj/164/orig -> origin/gh/fduwjj/164/orig 2025-07-17T08:34:50.7201315Z * [new branch] gh/fduwjj/165/base -> origin/gh/fduwjj/165/base 2025-07-17T08:34:50.7204654Z * [new branch] gh/fduwjj/165/head -> origin/gh/fduwjj/165/head 2025-07-17T08:34:50.7207997Z * [new branch] gh/fduwjj/165/orig -> origin/gh/fduwjj/165/orig 2025-07-17T08:34:50.7212627Z * [new branch] gh/fduwjj/166/base -> origin/gh/fduwjj/166/base 2025-07-17T08:34:50.7215921Z * [new branch] gh/fduwjj/166/head -> origin/gh/fduwjj/166/head 2025-07-17T08:34:50.7219358Z * [new branch] gh/fduwjj/166/orig -> origin/gh/fduwjj/166/orig 2025-07-17T08:34:50.7223869Z * [new branch] gh/fduwjj/167/base -> origin/gh/fduwjj/167/base 2025-07-17T08:34:50.7227313Z * [new branch] gh/fduwjj/167/head -> origin/gh/fduwjj/167/head 2025-07-17T08:34:50.7230633Z * [new branch] gh/fduwjj/167/orig -> origin/gh/fduwjj/167/orig 2025-07-17T08:34:50.7235283Z * [new branch] gh/fduwjj/168/base -> origin/gh/fduwjj/168/base 2025-07-17T08:34:50.7238679Z * [new branch] gh/fduwjj/168/head -> origin/gh/fduwjj/168/head 2025-07-17T08:34:50.7242039Z * [new branch] gh/fduwjj/168/orig -> origin/gh/fduwjj/168/orig 2025-07-17T08:34:50.7246714Z * [new branch] gh/fduwjj/169/base -> origin/gh/fduwjj/169/base 2025-07-17T08:34:50.7249890Z * [new branch] gh/fduwjj/169/head -> origin/gh/fduwjj/169/head 2025-07-17T08:34:50.7253450Z * [new branch] gh/fduwjj/169/orig -> origin/gh/fduwjj/169/orig 2025-07-17T08:34:50.7258894Z * [new branch] gh/fegin/306/base -> origin/gh/fegin/306/base 2025-07-17T08:34:50.7262182Z * [new branch] gh/fegin/306/head -> origin/gh/fegin/306/head 2025-07-17T08:34:50.7265527Z * [new branch] gh/fegin/306/orig -> origin/gh/fegin/306/orig 2025-07-17T08:34:50.7270169Z * [new branch] gh/fegin/307/base -> origin/gh/fegin/307/base 2025-07-17T08:34:50.7273529Z * [new branch] gh/fegin/307/head -> origin/gh/fegin/307/head 2025-07-17T08:34:50.7277111Z * [new branch] gh/fegin/307/orig -> origin/gh/fegin/307/orig 2025-07-17T08:34:50.7282426Z * [new branch] gh/fffrog/103/base -> origin/gh/fffrog/103/base 2025-07-17T08:34:50.7285725Z * [new branch] gh/fffrog/103/head -> origin/gh/fffrog/103/head 2025-07-17T08:34:50.7289164Z * [new branch] gh/fffrog/103/orig -> origin/gh/fffrog/103/orig 2025-07-17T08:34:50.7293721Z * [new branch] gh/fffrog/104/base -> origin/gh/fffrog/104/base 2025-07-17T08:34:50.7297048Z * [new branch] gh/fffrog/104/head -> origin/gh/fffrog/104/head 2025-07-17T08:34:50.7300441Z * [new branch] gh/fffrog/104/orig -> origin/gh/fffrog/104/orig 2025-07-17T08:34:50.7304902Z * [new branch] gh/fffrog/105/base -> origin/gh/fffrog/105/base 2025-07-17T08:34:50.7308266Z * [new branch] gh/fffrog/105/head -> origin/gh/fffrog/105/head 2025-07-17T08:34:50.7311586Z * [new branch] gh/fffrog/105/orig -> origin/gh/fffrog/105/orig 2025-07-17T08:34:50.7316458Z * [new branch] gh/fffrog/106/base -> origin/gh/fffrog/106/base 2025-07-17T08:34:50.7319730Z * [new branch] gh/fffrog/106/head -> origin/gh/fffrog/106/head 2025-07-17T08:34:50.7323084Z * [new branch] gh/fffrog/106/orig -> origin/gh/fffrog/106/orig 2025-07-17T08:34:50.7327557Z * [new branch] gh/fffrog/107/base -> origin/gh/fffrog/107/base 2025-07-17T08:34:50.7330889Z * [new branch] gh/fffrog/107/head -> origin/gh/fffrog/107/head 2025-07-17T08:34:50.7334196Z * [new branch] gh/fffrog/107/orig -> origin/gh/fffrog/107/orig 2025-07-17T08:34:50.7338730Z * [new branch] gh/fffrog/108/base -> origin/gh/fffrog/108/base 2025-07-17T08:34:50.7342085Z * [new branch] gh/fffrog/108/head -> origin/gh/fffrog/108/head 2025-07-17T08:34:50.7355803Z * [new branch] gh/fffrog/108/orig -> origin/gh/fffrog/108/orig 2025-07-17T08:34:50.7356253Z * [new branch] gh/fffrog/109/base -> origin/gh/fffrog/109/base 2025-07-17T08:34:50.7356572Z * [new branch] gh/fffrog/109/head -> origin/gh/fffrog/109/head 2025-07-17T08:34:50.7356884Z * [new branch] gh/fffrog/109/orig -> origin/gh/fffrog/109/orig 2025-07-17T08:34:50.7361449Z * [new branch] gh/fffrog/110/base -> origin/gh/fffrog/110/base 2025-07-17T08:34:50.7364930Z * [new branch] gh/fffrog/110/head -> origin/gh/fffrog/110/head 2025-07-17T08:34:50.7368250Z * [new branch] gh/fffrog/110/orig -> origin/gh/fffrog/110/orig 2025-07-17T08:34:50.7372796Z * [new branch] gh/fffrog/111/base -> origin/gh/fffrog/111/base 2025-07-17T08:34:50.7376027Z * [new branch] gh/fffrog/111/head -> origin/gh/fffrog/111/head 2025-07-17T08:34:50.7379336Z * [new branch] gh/fffrog/111/orig -> origin/gh/fffrog/111/orig 2025-07-17T08:34:50.7384061Z * [new branch] gh/fffrog/112/base -> origin/gh/fffrog/112/base 2025-07-17T08:34:50.7387155Z * [new branch] gh/fffrog/112/head -> origin/gh/fffrog/112/head 2025-07-17T08:34:50.7390442Z * [new branch] gh/fffrog/112/orig -> origin/gh/fffrog/112/orig 2025-07-17T08:34:50.7395001Z * [new branch] gh/fffrog/113/base -> origin/gh/fffrog/113/base 2025-07-17T08:34:50.7398344Z * [new branch] gh/fffrog/113/head -> origin/gh/fffrog/113/head 2025-07-17T08:34:50.7401823Z * [new branch] gh/fffrog/113/orig -> origin/gh/fffrog/113/orig 2025-07-17T08:34:50.7406426Z * [new branch] gh/fffrog/114/base -> origin/gh/fffrog/114/base 2025-07-17T08:34:50.7409695Z * [new branch] gh/fffrog/114/head -> origin/gh/fffrog/114/head 2025-07-17T08:34:50.7413151Z * [new branch] gh/fffrog/114/orig -> origin/gh/fffrog/114/orig 2025-07-17T08:34:50.7417659Z * [new branch] gh/fffrog/115/base -> origin/gh/fffrog/115/base 2025-07-17T08:34:50.7420931Z * [new branch] gh/fffrog/115/head -> origin/gh/fffrog/115/head 2025-07-17T08:34:50.7424246Z * [new branch] gh/fffrog/115/orig -> origin/gh/fffrog/115/orig 2025-07-17T08:34:50.7428796Z * [new branch] gh/fffrog/116/base -> origin/gh/fffrog/116/base 2025-07-17T08:34:50.7432131Z * [new branch] gh/fffrog/116/head -> origin/gh/fffrog/116/head 2025-07-17T08:34:50.7435704Z * [new branch] gh/fffrog/116/orig -> origin/gh/fffrog/116/orig 2025-07-17T08:34:50.7440135Z * [new branch] gh/fffrog/117/base -> origin/gh/fffrog/117/base 2025-07-17T08:34:50.7443650Z * [new branch] gh/fffrog/117/head -> origin/gh/fffrog/117/head 2025-07-17T08:34:50.7447229Z * [new branch] gh/fffrog/117/orig -> origin/gh/fffrog/117/orig 2025-07-17T08:34:50.7451975Z * [new branch] gh/fffrog/118/base -> origin/gh/fffrog/118/base 2025-07-17T08:34:50.7455373Z * [new branch] gh/fffrog/118/head -> origin/gh/fffrog/118/head 2025-07-17T08:34:50.7458751Z * [new branch] gh/fffrog/118/orig -> origin/gh/fffrog/118/orig 2025-07-17T08:34:50.7463287Z * [new branch] gh/fffrog/119/base -> origin/gh/fffrog/119/base 2025-07-17T08:34:50.7466625Z * [new branch] gh/fffrog/119/head -> origin/gh/fffrog/119/head 2025-07-17T08:34:50.7469983Z * [new branch] gh/fffrog/119/orig -> origin/gh/fffrog/119/orig 2025-07-17T08:34:50.7474578Z * [new branch] gh/fffrog/120/base -> origin/gh/fffrog/120/base 2025-07-17T08:34:50.7478030Z * [new branch] gh/fffrog/120/head -> origin/gh/fffrog/120/head 2025-07-17T08:34:50.7481273Z * [new branch] gh/fffrog/120/orig -> origin/gh/fffrog/120/orig 2025-07-17T08:34:50.7485737Z * [new branch] gh/fffrog/121/base -> origin/gh/fffrog/121/base 2025-07-17T08:34:50.7489056Z * [new branch] gh/fffrog/121/head -> origin/gh/fffrog/121/head 2025-07-17T08:34:50.7492366Z * [new branch] gh/fffrog/121/orig -> origin/gh/fffrog/121/orig 2025-07-17T08:34:50.7496933Z * [new branch] gh/fffrog/122/base -> origin/gh/fffrog/122/base 2025-07-17T08:34:50.7500318Z * [new branch] gh/fffrog/122/head -> origin/gh/fffrog/122/head 2025-07-17T08:34:50.7503631Z * [new branch] gh/fffrog/122/orig -> origin/gh/fffrog/122/orig 2025-07-17T08:34:50.7508226Z * [new branch] gh/fffrog/123/base -> origin/gh/fffrog/123/base 2025-07-17T08:34:50.7511550Z * [new branch] gh/fffrog/123/head -> origin/gh/fffrog/123/head 2025-07-17T08:34:50.7515374Z * [new branch] gh/fffrog/123/orig -> origin/gh/fffrog/123/orig 2025-07-17T08:34:50.7519782Z * [new branch] gh/fffrog/39/base -> origin/gh/fffrog/39/base 2025-07-17T08:34:50.7523124Z * [new branch] gh/fffrog/39/head -> origin/gh/fffrog/39/head 2025-07-17T08:34:50.7526496Z * [new branch] gh/fffrog/39/orig -> origin/gh/fffrog/39/orig 2025-07-17T08:34:50.7530967Z * [new branch] gh/fffrog/67/base -> origin/gh/fffrog/67/base 2025-07-17T08:34:50.7534262Z * [new branch] gh/fffrog/67/head -> origin/gh/fffrog/67/head 2025-07-17T08:34:50.7537587Z * [new branch] gh/fffrog/67/orig -> origin/gh/fffrog/67/orig 2025-07-17T08:34:50.7542359Z * [new branch] gh/fffrog/87/base -> origin/gh/fffrog/87/base 2025-07-17T08:34:50.7545786Z * [new branch] gh/fffrog/87/head -> origin/gh/fffrog/87/head 2025-07-17T08:34:50.7549134Z * [new branch] gh/fffrog/87/orig -> origin/gh/fffrog/87/orig 2025-07-17T08:34:50.7553726Z * [new branch] gh/fffrog/93/base -> origin/gh/fffrog/93/base 2025-07-17T08:34:50.7557302Z * [new branch] gh/fffrog/93/head -> origin/gh/fffrog/93/head 2025-07-17T08:34:50.7560576Z * [new branch] gh/fffrog/93/orig -> origin/gh/fffrog/93/orig 2025-07-17T08:34:50.7565123Z * [new branch] gh/fffrog/99/base -> origin/gh/fffrog/99/base 2025-07-17T08:34:50.7568458Z * [new branch] gh/fffrog/99/head -> origin/gh/fffrog/99/head 2025-07-17T08:34:50.7571767Z * [new branch] gh/fffrog/99/orig -> origin/gh/fffrog/99/orig 2025-07-17T08:34:50.7577290Z * [new branch] gh/gmagogsfm/1/base -> origin/gh/gmagogsfm/1/base 2025-07-17T08:34:50.7580605Z * [new branch] gh/gmagogsfm/1/head -> origin/gh/gmagogsfm/1/head 2025-07-17T08:34:50.7584120Z * [new branch] gh/gmagogsfm/1/orig -> origin/gh/gmagogsfm/1/orig 2025-07-17T08:34:50.7588515Z * [new branch] gh/gmagogsfm/2/base -> origin/gh/gmagogsfm/2/base 2025-07-17T08:34:50.7591803Z * [new branch] gh/gmagogsfm/2/head -> origin/gh/gmagogsfm/2/head 2025-07-17T08:34:50.7595725Z * [new branch] gh/gmagogsfm/2/orig -> origin/gh/gmagogsfm/2/orig 2025-07-17T08:34:50.7599787Z * [new branch] gh/gmagogsfm/3/base -> origin/gh/gmagogsfm/3/base 2025-07-17T08:34:50.7603109Z * [new branch] gh/gmagogsfm/3/head -> origin/gh/gmagogsfm/3/head 2025-07-17T08:34:50.7606422Z * [new branch] gh/gmagogsfm/3/orig -> origin/gh/gmagogsfm/3/orig 2025-07-17T08:34:50.7610976Z * [new branch] gh/gmagogsfm/4/base -> origin/gh/gmagogsfm/4/base 2025-07-17T08:34:50.7614301Z * [new branch] gh/gmagogsfm/4/head -> origin/gh/gmagogsfm/4/head 2025-07-17T08:34:50.7617635Z * [new branch] gh/gmagogsfm/4/orig -> origin/gh/gmagogsfm/4/orig 2025-07-17T08:34:50.7623237Z * [new branch] gh/guangyey/130/base -> origin/gh/guangyey/130/base 2025-07-17T08:34:50.7626679Z * [new branch] gh/guangyey/130/head -> origin/gh/guangyey/130/head 2025-07-17T08:34:50.7630007Z * [new branch] gh/guangyey/130/orig -> origin/gh/guangyey/130/orig 2025-07-17T08:34:50.7634849Z * [new branch] gh/guangyey/133/base -> origin/gh/guangyey/133/base 2025-07-17T08:34:50.7638276Z * [new branch] gh/guangyey/133/head -> origin/gh/guangyey/133/head 2025-07-17T08:34:50.7641602Z * [new branch] gh/guangyey/133/orig -> origin/gh/guangyey/133/orig 2025-07-17T08:34:50.7646181Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-07-17T08:34:50.7649513Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-07-17T08:34:50.7653321Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-07-17T08:34:50.7657629Z * [new branch] gh/guangyey/135/base -> origin/gh/guangyey/135/base 2025-07-17T08:34:50.7661041Z * [new branch] gh/guangyey/135/head -> origin/gh/guangyey/135/head 2025-07-17T08:34:50.7664376Z * [new branch] gh/guangyey/135/orig -> origin/gh/guangyey/135/orig 2025-07-17T08:34:50.7668914Z * [new branch] gh/guangyey/139/base -> origin/gh/guangyey/139/base 2025-07-17T08:34:50.7672317Z * [new branch] gh/guangyey/139/head -> origin/gh/guangyey/139/head 2025-07-17T08:34:50.7675892Z * [new branch] gh/guangyey/139/orig -> origin/gh/guangyey/139/orig 2025-07-17T08:34:50.7680358Z * [new branch] gh/guangyey/140/base -> origin/gh/guangyey/140/base 2025-07-17T08:34:50.7683688Z * [new branch] gh/guangyey/140/head -> origin/gh/guangyey/140/head 2025-07-17T08:34:50.7686985Z * [new branch] gh/guangyey/140/orig -> origin/gh/guangyey/140/orig 2025-07-17T08:34:50.7691543Z * [new branch] gh/guangyey/142/base -> origin/gh/guangyey/142/base 2025-07-17T08:34:50.7694828Z * [new branch] gh/guangyey/142/head -> origin/gh/guangyey/142/head 2025-07-17T08:34:50.7698201Z * [new branch] gh/guangyey/142/orig -> origin/gh/guangyey/142/orig 2025-07-17T08:34:50.7702783Z * [new branch] gh/guangyey/145/base -> origin/gh/guangyey/145/base 2025-07-17T08:34:50.7706093Z * [new branch] gh/guangyey/145/head -> origin/gh/guangyey/145/head 2025-07-17T08:34:50.7709434Z * [new branch] gh/guangyey/145/orig -> origin/gh/guangyey/145/orig 2025-07-17T08:34:50.7714003Z * [new branch] gh/guangyey/153/base -> origin/gh/guangyey/153/base 2025-07-17T08:34:50.7717571Z * [new branch] gh/guangyey/153/head -> origin/gh/guangyey/153/head 2025-07-17T08:34:50.7720881Z * [new branch] gh/guangyey/153/orig -> origin/gh/guangyey/153/orig 2025-07-17T08:34:50.7725387Z * [new branch] gh/guangyey/154/base -> origin/gh/guangyey/154/base 2025-07-17T08:34:50.7728742Z * [new branch] gh/guangyey/154/head -> origin/gh/guangyey/154/head 2025-07-17T08:34:50.7732110Z * [new branch] gh/guangyey/154/orig -> origin/gh/guangyey/154/orig 2025-07-17T08:34:50.7736640Z * [new branch] gh/guangyey/158/base -> origin/gh/guangyey/158/base 2025-07-17T08:34:50.7740011Z * [new branch] gh/guangyey/158/head -> origin/gh/guangyey/158/head 2025-07-17T08:34:50.7743349Z * [new branch] gh/guangyey/158/orig -> origin/gh/guangyey/158/orig 2025-07-17T08:34:50.7747899Z * [new branch] gh/guangyey/159/base -> origin/gh/guangyey/159/base 2025-07-17T08:34:50.7751242Z * [new branch] gh/guangyey/159/head -> origin/gh/guangyey/159/head 2025-07-17T08:34:50.7754683Z * [new branch] gh/guangyey/159/orig -> origin/gh/guangyey/159/orig 2025-07-17T08:34:50.7759699Z * [new branch] gh/guangyey/160/base -> origin/gh/guangyey/160/base 2025-07-17T08:34:50.7763117Z * [new branch] gh/guangyey/160/head -> origin/gh/guangyey/160/head 2025-07-17T08:34:50.7766563Z * [new branch] gh/guangyey/160/orig -> origin/gh/guangyey/160/orig 2025-07-17T08:34:50.7808067Z * [new branch] gh/guangyey/161/base -> origin/gh/guangyey/161/base 2025-07-17T08:34:50.8005585Z * [new branch] gh/guangyey/161/head -> origin/gh/guangyey/161/head 2025-07-17T08:34:50.8053180Z * [new branch] gh/guangyey/161/orig -> origin/gh/guangyey/161/orig 2025-07-17T08:34:50.8053928Z * [new branch] gh/guangyey/162/base -> origin/gh/guangyey/162/base 2025-07-17T08:34:50.8055112Z * [new branch] gh/guangyey/162/head -> origin/gh/guangyey/162/head 2025-07-17T08:34:50.8055548Z * [new branch] gh/guangyey/162/orig -> origin/gh/guangyey/162/orig 2025-07-17T08:34:50.8055938Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-07-17T08:34:50.8056307Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-07-17T08:34:50.8056720Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-07-17T08:34:50.8057355Z * [new branch] gh/guangyey/164/base -> origin/gh/guangyey/164/base 2025-07-17T08:34:50.8057734Z * [new branch] gh/guangyey/164/head -> origin/gh/guangyey/164/head 2025-07-17T08:34:50.8058108Z * [new branch] gh/guangyey/164/orig -> origin/gh/guangyey/164/orig 2025-07-17T08:34:50.8058473Z * [new branch] gh/guangyey/165/base -> origin/gh/guangyey/165/base 2025-07-17T08:34:50.8058858Z * [new branch] gh/guangyey/165/head -> origin/gh/guangyey/165/head 2025-07-17T08:34:50.8059226Z * [new branch] gh/guangyey/165/orig -> origin/gh/guangyey/165/orig 2025-07-17T08:34:50.8059595Z * [new branch] gh/guangyey/166/base -> origin/gh/guangyey/166/base 2025-07-17T08:34:50.8059970Z * [new branch] gh/guangyey/166/head -> origin/gh/guangyey/166/head 2025-07-17T08:34:50.8060339Z * [new branch] gh/guangyey/166/orig -> origin/gh/guangyey/166/orig 2025-07-17T08:34:50.8060721Z * [new branch] gh/guangyey/167/base -> origin/gh/guangyey/167/base 2025-07-17T08:34:50.8061086Z * [new branch] gh/guangyey/167/head -> origin/gh/guangyey/167/head 2025-07-17T08:34:50.8061459Z * [new branch] gh/guangyey/167/orig -> origin/gh/guangyey/167/orig 2025-07-17T08:34:50.8061837Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-07-17T08:34:50.8062203Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-07-17T08:34:50.8062575Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-07-17T08:34:50.8062946Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-07-17T08:34:50.8063323Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-07-17T08:34:50.8063698Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-07-17T08:34:50.8064067Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-07-17T08:34:50.8064444Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-07-17T08:34:50.8064861Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-07-17T08:34:50.8065243Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-07-17T08:34:50.8065620Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-07-17T08:34:50.8065984Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-07-17T08:34:50.8066357Z * [new branch] gh/guangyey/172/base -> origin/gh/guangyey/172/base 2025-07-17T08:34:50.8066723Z * [new branch] gh/guangyey/172/head -> origin/gh/guangyey/172/head 2025-07-17T08:34:50.8067102Z * [new branch] gh/guangyey/172/orig -> origin/gh/guangyey/172/orig 2025-07-17T08:34:50.8067479Z * [new branch] gh/guangyey/79/base -> origin/gh/guangyey/79/base 2025-07-17T08:34:50.8067858Z * [new branch] gh/guangyey/79/head -> origin/gh/guangyey/79/head 2025-07-17T08:34:50.8068231Z * [new branch] gh/guangyey/79/orig -> origin/gh/guangyey/79/orig 2025-07-17T08:34:50.8068724Z * [new branch] gh/guangyey/89/base -> origin/gh/guangyey/89/base 2025-07-17T08:34:50.8069089Z * [new branch] gh/guangyey/89/head -> origin/gh/guangyey/89/head 2025-07-17T08:34:50.8069451Z * [new branch] gh/guangyey/89/orig -> origin/gh/guangyey/89/orig 2025-07-17T08:34:50.8069856Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-07-17T08:34:50.8070302Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-07-17T08:34:50.8070893Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-07-17T08:34:50.8071331Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-07-17T08:34:50.8071763Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-07-17T08:34:50.8072199Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-07-17T08:34:50.8072636Z * [new branch] gh/guilhermeleobas/120/base -> origin/gh/guilhermeleobas/120/base 2025-07-17T08:34:50.8073059Z * [new branch] gh/guilhermeleobas/120/head -> origin/gh/guilhermeleobas/120/head 2025-07-17T08:34:50.8073493Z * [new branch] gh/guilhermeleobas/120/orig -> origin/gh/guilhermeleobas/120/orig 2025-07-17T08:34:50.8073992Z * [new branch] gh/guilhermeleobas/124/base -> origin/gh/guilhermeleobas/124/base 2025-07-17T08:34:50.8074434Z * [new branch] gh/guilhermeleobas/124/head -> origin/gh/guilhermeleobas/124/head 2025-07-17T08:34:50.8074869Z * [new branch] gh/guilhermeleobas/124/orig -> origin/gh/guilhermeleobas/124/orig 2025-07-17T08:34:50.8075333Z * [new branch] gh/guilhermeleobas/127/base -> origin/gh/guilhermeleobas/127/base 2025-07-17T08:34:50.8075763Z * [new branch] gh/guilhermeleobas/127/head -> origin/gh/guilhermeleobas/127/head 2025-07-17T08:34:50.8076190Z * [new branch] gh/guilhermeleobas/127/orig -> origin/gh/guilhermeleobas/127/orig 2025-07-17T08:34:50.8076620Z * [new branch] gh/guilhermeleobas/143/base -> origin/gh/guilhermeleobas/143/base 2025-07-17T08:34:50.8077057Z * [new branch] gh/guilhermeleobas/143/head -> origin/gh/guilhermeleobas/143/head 2025-07-17T08:34:50.8077479Z * [new branch] gh/guilhermeleobas/143/orig -> origin/gh/guilhermeleobas/143/orig 2025-07-17T08:34:50.8077910Z * [new branch] gh/guilhermeleobas/145/base -> origin/gh/guilhermeleobas/145/base 2025-07-17T08:34:50.8078335Z * [new branch] gh/guilhermeleobas/145/head -> origin/gh/guilhermeleobas/145/head 2025-07-17T08:34:50.8078769Z * [new branch] gh/guilhermeleobas/145/orig -> origin/gh/guilhermeleobas/145/orig 2025-07-17T08:34:50.8079200Z * [new branch] gh/guilhermeleobas/146/base -> origin/gh/guilhermeleobas/146/base 2025-07-17T08:34:50.8079629Z * [new branch] gh/guilhermeleobas/146/head -> origin/gh/guilhermeleobas/146/head 2025-07-17T08:34:50.8080064Z * [new branch] gh/guilhermeleobas/146/orig -> origin/gh/guilhermeleobas/146/orig 2025-07-17T08:34:50.8080488Z * [new branch] gh/guilhermeleobas/147/base -> origin/gh/guilhermeleobas/147/base 2025-07-17T08:34:50.8080930Z * [new branch] gh/guilhermeleobas/147/head -> origin/gh/guilhermeleobas/147/head 2025-07-17T08:34:50.8081362Z * [new branch] gh/guilhermeleobas/147/orig -> origin/gh/guilhermeleobas/147/orig 2025-07-17T08:34:50.8081798Z * [new branch] gh/guilhermeleobas/148/base -> origin/gh/guilhermeleobas/148/base 2025-07-17T08:34:50.8082230Z * [new branch] gh/guilhermeleobas/148/head -> origin/gh/guilhermeleobas/148/head 2025-07-17T08:34:50.8090390Z * [new branch] gh/guilhermeleobas/148/orig -> origin/gh/guilhermeleobas/148/orig 2025-07-17T08:34:50.8091339Z * [new branch] gh/guilhermeleobas/149/base -> origin/gh/guilhermeleobas/149/base 2025-07-17T08:34:50.8091949Z * [new branch] gh/guilhermeleobas/149/head -> origin/gh/guilhermeleobas/149/head 2025-07-17T08:34:50.8092633Z * [new branch] gh/guilhermeleobas/149/orig -> origin/gh/guilhermeleobas/149/orig 2025-07-17T08:34:50.8095244Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-07-17T08:34:50.8114103Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-07-17T08:34:50.8114900Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-07-17T08:34:50.8115470Z * [new branch] gh/guilhermeleobas/151/base -> origin/gh/guilhermeleobas/151/base 2025-07-17T08:34:50.8116095Z * [new branch] gh/guilhermeleobas/151/head -> origin/gh/guilhermeleobas/151/head 2025-07-17T08:34:50.8117067Z * [new branch] gh/guilhermeleobas/151/orig -> origin/gh/guilhermeleobas/151/orig 2025-07-17T08:34:50.8152983Z * [new branch] gh/guilhermeleobas/152/base -> origin/gh/guilhermeleobas/152/base 2025-07-17T08:34:50.8181067Z * [new branch] gh/guilhermeleobas/152/head -> origin/gh/guilhermeleobas/152/head 2025-07-17T08:34:50.8181574Z * [new branch] gh/guilhermeleobas/152/orig -> origin/gh/guilhermeleobas/152/orig 2025-07-17T08:34:50.8186883Z * [new branch] gh/guilhermeleobas/153/base -> origin/gh/guilhermeleobas/153/base 2025-07-17T08:34:50.8236152Z * [new branch] gh/guilhermeleobas/153/head -> origin/gh/guilhermeleobas/153/head 2025-07-17T08:34:50.8236742Z * [new branch] gh/guilhermeleobas/153/orig -> origin/gh/guilhermeleobas/153/orig 2025-07-17T08:34:50.8237234Z * [new branch] gh/guilhermeleobas/154/base -> origin/gh/guilhermeleobas/154/base 2025-07-17T08:34:50.8237688Z * [new branch] gh/guilhermeleobas/154/head -> origin/gh/guilhermeleobas/154/head 2025-07-17T08:34:50.8238188Z * [new branch] gh/guilhermeleobas/154/orig -> origin/gh/guilhermeleobas/154/orig 2025-07-17T08:34:50.8238635Z * [new branch] gh/guilhermeleobas/155/base -> origin/gh/guilhermeleobas/155/base 2025-07-17T08:34:50.8239072Z * [new branch] gh/guilhermeleobas/155/head -> origin/gh/guilhermeleobas/155/head 2025-07-17T08:34:50.8239511Z * [new branch] gh/guilhermeleobas/155/orig -> origin/gh/guilhermeleobas/155/orig 2025-07-17T08:34:50.8239941Z * [new branch] gh/guilhermeleobas/156/base -> origin/gh/guilhermeleobas/156/base 2025-07-17T08:34:50.8240409Z * [new branch] gh/guilhermeleobas/156/head -> origin/gh/guilhermeleobas/156/head 2025-07-17T08:34:50.8240848Z * [new branch] gh/guilhermeleobas/156/orig -> origin/gh/guilhermeleobas/156/orig 2025-07-17T08:34:50.8241275Z * [new branch] gh/guilhermeleobas/157/base -> origin/gh/guilhermeleobas/157/base 2025-07-17T08:34:50.8241715Z * [new branch] gh/guilhermeleobas/157/head -> origin/gh/guilhermeleobas/157/head 2025-07-17T08:34:50.8242145Z * [new branch] gh/guilhermeleobas/157/orig -> origin/gh/guilhermeleobas/157/orig 2025-07-17T08:34:50.8242575Z * [new branch] gh/guilhermeleobas/158/base -> origin/gh/guilhermeleobas/158/base 2025-07-17T08:34:50.8243001Z * [new branch] gh/guilhermeleobas/158/head -> origin/gh/guilhermeleobas/158/head 2025-07-17T08:34:50.8243423Z * [new branch] gh/guilhermeleobas/158/orig -> origin/gh/guilhermeleobas/158/orig 2025-07-17T08:34:50.8243863Z * [new branch] gh/guilhermeleobas/159/base -> origin/gh/guilhermeleobas/159/base 2025-07-17T08:34:50.8244287Z * [new branch] gh/guilhermeleobas/159/head -> origin/gh/guilhermeleobas/159/head 2025-07-17T08:34:50.8244720Z * [new branch] gh/guilhermeleobas/159/orig -> origin/gh/guilhermeleobas/159/orig 2025-07-17T08:34:50.8245820Z * [new branch] gh/guilhermeleobas/160/base -> origin/gh/guilhermeleobas/160/base 2025-07-17T08:34:50.8246260Z * [new branch] gh/guilhermeleobas/160/head -> origin/gh/guilhermeleobas/160/head 2025-07-17T08:34:50.8246855Z * [new branch] gh/guilhermeleobas/160/orig -> origin/gh/guilhermeleobas/160/orig 2025-07-17T08:34:50.8247296Z * [new branch] gh/guilhermeleobas/161/base -> origin/gh/guilhermeleobas/161/base 2025-07-17T08:34:50.8247768Z * [new branch] gh/guilhermeleobas/161/head -> origin/gh/guilhermeleobas/161/head 2025-07-17T08:34:50.8252438Z * [new branch] gh/guilhermeleobas/161/orig -> origin/gh/guilhermeleobas/161/orig 2025-07-17T08:34:50.8263198Z * [new branch] gh/guilhermeleobas/162/base -> origin/gh/guilhermeleobas/162/base 2025-07-17T08:34:50.8264024Z * [new branch] gh/guilhermeleobas/162/head -> origin/gh/guilhermeleobas/162/head 2025-07-17T08:34:50.8266508Z * [new branch] gh/guilhermeleobas/162/orig -> origin/gh/guilhermeleobas/162/orig 2025-07-17T08:34:50.8272342Z * [new branch] gh/guilhermeleobas/163/base -> origin/gh/guilhermeleobas/163/base 2025-07-17T08:34:50.8280887Z * [new branch] gh/guilhermeleobas/163/head -> origin/gh/guilhermeleobas/163/head 2025-07-17T08:34:50.8295794Z * [new branch] gh/guilhermeleobas/163/orig -> origin/gh/guilhermeleobas/163/orig 2025-07-17T08:34:50.8301557Z * [new branch] gh/guilhermeleobas/164/base -> origin/gh/guilhermeleobas/164/base 2025-07-17T08:34:50.8302159Z * [new branch] gh/guilhermeleobas/164/head -> origin/gh/guilhermeleobas/164/head 2025-07-17T08:34:50.8302688Z * [new branch] gh/guilhermeleobas/164/orig -> origin/gh/guilhermeleobas/164/orig 2025-07-17T08:34:50.8303188Z * [new branch] gh/guilhermeleobas/165/base -> origin/gh/guilhermeleobas/165/base 2025-07-17T08:34:50.8304523Z * [new branch] gh/guilhermeleobas/165/head -> origin/gh/guilhermeleobas/165/head 2025-07-17T08:34:50.8308541Z * [new branch] gh/guilhermeleobas/165/orig -> origin/gh/guilhermeleobas/165/orig 2025-07-17T08:34:50.8314571Z * [new branch] gh/guilhermeleobas/166/base -> origin/gh/guilhermeleobas/166/base 2025-07-17T08:34:50.8317714Z * [new branch] gh/guilhermeleobas/166/head -> origin/gh/guilhermeleobas/166/head 2025-07-17T08:34:50.8324874Z * [new branch] gh/guilhermeleobas/166/orig -> origin/gh/guilhermeleobas/166/orig 2025-07-17T08:34:50.8327814Z * [new branch] gh/guilhermeleobas/167/base -> origin/gh/guilhermeleobas/167/base 2025-07-17T08:34:50.8333112Z * [new branch] gh/guilhermeleobas/167/head -> origin/gh/guilhermeleobas/167/head 2025-07-17T08:34:50.8348423Z * [new branch] gh/guilhermeleobas/167/orig -> origin/gh/guilhermeleobas/167/orig 2025-07-17T08:34:50.8350783Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-07-17T08:34:50.8351375Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-07-17T08:34:50.8351827Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-07-17T08:34:50.8352949Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-07-17T08:34:50.8356416Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-07-17T08:34:50.8360566Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-07-17T08:34:50.8367020Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-07-17T08:34:50.8369092Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-07-17T08:34:50.8372901Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-07-17T08:34:50.8377830Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-07-17T08:34:50.8380921Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-07-17T08:34:50.8384632Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-07-17T08:34:50.8389727Z * [new branch] gh/guilhermeleobas/172/base -> origin/gh/guilhermeleobas/172/base 2025-07-17T08:34:50.8393117Z * [new branch] gh/guilhermeleobas/172/head -> origin/gh/guilhermeleobas/172/head 2025-07-17T08:34:50.8396767Z * [new branch] gh/guilhermeleobas/172/orig -> origin/gh/guilhermeleobas/172/orig 2025-07-17T08:34:50.8401483Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-07-17T08:34:50.8404936Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-07-17T08:34:50.8408727Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-07-17T08:34:50.8414594Z * [new branch] gh/guilhermeleobas/174/base -> origin/gh/guilhermeleobas/174/base 2025-07-17T08:34:50.8418068Z * [new branch] gh/guilhermeleobas/174/head -> origin/gh/guilhermeleobas/174/head 2025-07-17T08:34:50.8421320Z * [new branch] gh/guilhermeleobas/174/orig -> origin/gh/guilhermeleobas/174/orig 2025-07-17T08:34:50.8425941Z * [new branch] gh/guilhermeleobas/175/base -> origin/gh/guilhermeleobas/175/base 2025-07-17T08:34:50.8429451Z * [new branch] gh/guilhermeleobas/175/head -> origin/gh/guilhermeleobas/175/head 2025-07-17T08:34:50.8433293Z * [new branch] gh/guilhermeleobas/175/orig -> origin/gh/guilhermeleobas/175/orig 2025-07-17T08:34:50.8438332Z * [new branch] gh/guilhermeleobas/176/base -> origin/gh/guilhermeleobas/176/base 2025-07-17T08:34:50.8441762Z * [new branch] gh/guilhermeleobas/176/head -> origin/gh/guilhermeleobas/176/head 2025-07-17T08:34:50.8445417Z * [new branch] gh/guilhermeleobas/176/orig -> origin/gh/guilhermeleobas/176/orig 2025-07-17T08:34:50.8450220Z * [new branch] gh/guilhermeleobas/177/base -> origin/gh/guilhermeleobas/177/base 2025-07-17T08:34:50.8454015Z * [new branch] gh/guilhermeleobas/177/head -> origin/gh/guilhermeleobas/177/head 2025-07-17T08:34:50.8457411Z * [new branch] gh/guilhermeleobas/177/orig -> origin/gh/guilhermeleobas/177/orig 2025-07-17T08:34:50.8462342Z * [new branch] gh/guilhermeleobas/178/base -> origin/gh/guilhermeleobas/178/base 2025-07-17T08:34:50.8465802Z * [new branch] gh/guilhermeleobas/178/head -> origin/gh/guilhermeleobas/178/head 2025-07-17T08:34:50.8469333Z * [new branch] gh/guilhermeleobas/178/orig -> origin/gh/guilhermeleobas/178/orig 2025-07-17T08:34:50.8474259Z * [new branch] gh/guilhermeleobas/179/base -> origin/gh/guilhermeleobas/179/base 2025-07-17T08:34:50.8478031Z * [new branch] gh/guilhermeleobas/179/head -> origin/gh/guilhermeleobas/179/head 2025-07-17T08:34:50.8481870Z * [new branch] gh/guilhermeleobas/179/orig -> origin/gh/guilhermeleobas/179/orig 2025-07-17T08:34:50.8486859Z * [new branch] gh/guilhermeleobas/180/base -> origin/gh/guilhermeleobas/180/base 2025-07-17T08:34:50.8490487Z * [new branch] gh/guilhermeleobas/180/head -> origin/gh/guilhermeleobas/180/head 2025-07-17T08:34:50.8496996Z * [new branch] gh/guilhermeleobas/180/orig -> origin/gh/guilhermeleobas/180/orig 2025-07-17T08:34:50.8499516Z * [new branch] gh/guilhermeleobas/181/base -> origin/gh/guilhermeleobas/181/base 2025-07-17T08:34:50.8512692Z * [new branch] gh/guilhermeleobas/181/head -> origin/gh/guilhermeleobas/181/head 2025-07-17T08:34:50.8541807Z * [new branch] gh/guilhermeleobas/181/orig -> origin/gh/guilhermeleobas/181/orig 2025-07-17T08:34:50.8543097Z * [new branch] gh/guilhermeleobas/182/base -> origin/gh/guilhermeleobas/182/base 2025-07-17T08:34:50.8543656Z * [new branch] gh/guilhermeleobas/182/head -> origin/gh/guilhermeleobas/182/head 2025-07-17T08:34:50.8544097Z * [new branch] gh/guilhermeleobas/182/orig -> origin/gh/guilhermeleobas/182/orig 2025-07-17T08:34:50.8544567Z * [new branch] gh/guilhermeleobas/183/base -> origin/gh/guilhermeleobas/183/base 2025-07-17T08:34:50.8545021Z * [new branch] gh/guilhermeleobas/183/head -> origin/gh/guilhermeleobas/183/head 2025-07-17T08:34:50.8545728Z * [new branch] gh/guilhermeleobas/183/orig -> origin/gh/guilhermeleobas/183/orig 2025-07-17T08:34:50.8546182Z * [new branch] gh/guilhermeleobas/184/base -> origin/gh/guilhermeleobas/184/base 2025-07-17T08:34:50.8546616Z * [new branch] gh/guilhermeleobas/184/head -> origin/gh/guilhermeleobas/184/head 2025-07-17T08:34:50.8547078Z * [new branch] gh/guilhermeleobas/184/orig -> origin/gh/guilhermeleobas/184/orig 2025-07-17T08:34:50.8549682Z * [new branch] gh/guilhermeleobas/185/base -> origin/gh/guilhermeleobas/185/base 2025-07-17T08:34:50.8552597Z * [new branch] gh/guilhermeleobas/185/head -> origin/gh/guilhermeleobas/185/head 2025-07-17T08:34:50.8556718Z * [new branch] gh/guilhermeleobas/185/orig -> origin/gh/guilhermeleobas/185/orig 2025-07-17T08:34:50.8561950Z * [new branch] gh/guilhermeleobas/186/base -> origin/gh/guilhermeleobas/186/base 2025-07-17T08:34:50.8566046Z * [new branch] gh/guilhermeleobas/186/head -> origin/gh/guilhermeleobas/186/head 2025-07-17T08:34:50.8569508Z * [new branch] gh/guilhermeleobas/186/orig -> origin/gh/guilhermeleobas/186/orig 2025-07-17T08:34:50.8574528Z * [new branch] gh/guilhermeleobas/187/base -> origin/gh/guilhermeleobas/187/base 2025-07-17T08:34:50.8577953Z * [new branch] gh/guilhermeleobas/187/head -> origin/gh/guilhermeleobas/187/head 2025-07-17T08:34:50.8581734Z * [new branch] gh/guilhermeleobas/187/orig -> origin/gh/guilhermeleobas/187/orig 2025-07-17T08:34:50.8586901Z * [new branch] gh/guilhermeleobas/188/base -> origin/gh/guilhermeleobas/188/base 2025-07-17T08:34:50.8590347Z * [new branch] gh/guilhermeleobas/188/head -> origin/gh/guilhermeleobas/188/head 2025-07-17T08:34:50.8594237Z * [new branch] gh/guilhermeleobas/188/orig -> origin/gh/guilhermeleobas/188/orig 2025-07-17T08:34:50.8598990Z * [new branch] gh/guilhermeleobas/189/base -> origin/gh/guilhermeleobas/189/base 2025-07-17T08:34:50.8602787Z * [new branch] gh/guilhermeleobas/189/head -> origin/gh/guilhermeleobas/189/head 2025-07-17T08:34:50.8606207Z * [new branch] gh/guilhermeleobas/189/orig -> origin/gh/guilhermeleobas/189/orig 2025-07-17T08:34:50.8611374Z * [new branch] gh/guilhermeleobas/190/base -> origin/gh/guilhermeleobas/190/base 2025-07-17T08:34:50.8615219Z * [new branch] gh/guilhermeleobas/190/head -> origin/gh/guilhermeleobas/190/head 2025-07-17T08:34:50.8618666Z * [new branch] gh/guilhermeleobas/190/orig -> origin/gh/guilhermeleobas/190/orig 2025-07-17T08:34:50.8623951Z * [new branch] gh/guilhermeleobas/191/base -> origin/gh/guilhermeleobas/191/base 2025-07-17T08:34:50.8627529Z * [new branch] gh/guilhermeleobas/191/head -> origin/gh/guilhermeleobas/191/head 2025-07-17T08:34:50.8631058Z * [new branch] gh/guilhermeleobas/191/orig -> origin/gh/guilhermeleobas/191/orig 2025-07-17T08:34:50.8636361Z * [new branch] gh/guilhermeleobas/192/base -> origin/gh/guilhermeleobas/192/base 2025-07-17T08:34:50.8640294Z * [new branch] gh/guilhermeleobas/192/head -> origin/gh/guilhermeleobas/192/head 2025-07-17T08:34:50.8643870Z * [new branch] gh/guilhermeleobas/192/orig -> origin/gh/guilhermeleobas/192/orig 2025-07-17T08:34:50.8649063Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-07-17T08:34:50.8652446Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-07-17T08:34:50.8655923Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-07-17T08:34:50.8660654Z * [new branch] gh/guilhermeleobas/194/base -> origin/gh/guilhermeleobas/194/base 2025-07-17T08:34:50.8664097Z * [new branch] gh/guilhermeleobas/194/head -> origin/gh/guilhermeleobas/194/head 2025-07-17T08:34:50.8667473Z * [new branch] gh/guilhermeleobas/194/orig -> origin/gh/guilhermeleobas/194/orig 2025-07-17T08:34:50.8672139Z * [new branch] gh/guilhermeleobas/195/base -> origin/gh/guilhermeleobas/195/base 2025-07-17T08:34:50.8676291Z * [new branch] gh/guilhermeleobas/195/head -> origin/gh/guilhermeleobas/195/head 2025-07-17T08:34:50.8680266Z * [new branch] gh/guilhermeleobas/195/orig -> origin/gh/guilhermeleobas/195/orig 2025-07-17T08:34:50.8684930Z * [new branch] gh/guilhermeleobas/196/base -> origin/gh/guilhermeleobas/196/base 2025-07-17T08:34:50.8688541Z * [new branch] gh/guilhermeleobas/196/head -> origin/gh/guilhermeleobas/196/head 2025-07-17T08:34:50.8691944Z * [new branch] gh/guilhermeleobas/196/orig -> origin/gh/guilhermeleobas/196/orig 2025-07-17T08:34:50.8696478Z * [new branch] gh/guilhermeleobas/197/base -> origin/gh/guilhermeleobas/197/base 2025-07-17T08:34:50.8699789Z * [new branch] gh/guilhermeleobas/197/head -> origin/gh/guilhermeleobas/197/head 2025-07-17T08:34:50.8703186Z * [new branch] gh/guilhermeleobas/197/orig -> origin/gh/guilhermeleobas/197/orig 2025-07-17T08:34:50.8707809Z * [new branch] gh/guilhermeleobas/198/base -> origin/gh/guilhermeleobas/198/base 2025-07-17T08:34:50.8711220Z * [new branch] gh/guilhermeleobas/198/head -> origin/gh/guilhermeleobas/198/head 2025-07-17T08:34:50.8714776Z * [new branch] gh/guilhermeleobas/198/orig -> origin/gh/guilhermeleobas/198/orig 2025-07-17T08:34:50.8719360Z * [new branch] gh/guilhermeleobas/199/base -> origin/gh/guilhermeleobas/199/base 2025-07-17T08:34:50.8722718Z * [new branch] gh/guilhermeleobas/199/head -> origin/gh/guilhermeleobas/199/head 2025-07-17T08:34:50.8726177Z * [new branch] gh/guilhermeleobas/199/orig -> origin/gh/guilhermeleobas/199/orig 2025-07-17T08:34:50.8730891Z * [new branch] gh/guilhermeleobas/200/base -> origin/gh/guilhermeleobas/200/base 2025-07-17T08:34:50.8734157Z * [new branch] gh/guilhermeleobas/200/head -> origin/gh/guilhermeleobas/200/head 2025-07-17T08:34:50.8737492Z * [new branch] gh/guilhermeleobas/200/orig -> origin/gh/guilhermeleobas/200/orig 2025-07-17T08:34:50.8742041Z * [new branch] gh/guilhermeleobas/201/base -> origin/gh/guilhermeleobas/201/base 2025-07-17T08:34:50.8745549Z * [new branch] gh/guilhermeleobas/201/head -> origin/gh/guilhermeleobas/201/head 2025-07-17T08:34:50.8748929Z * [new branch] gh/guilhermeleobas/201/orig -> origin/gh/guilhermeleobas/201/orig 2025-07-17T08:34:50.8753614Z * [new branch] gh/guilhermeleobas/202/base -> origin/gh/guilhermeleobas/202/base 2025-07-17T08:34:50.8757194Z * [new branch] gh/guilhermeleobas/202/head -> origin/gh/guilhermeleobas/202/head 2025-07-17T08:34:50.8760665Z * [new branch] gh/guilhermeleobas/202/orig -> origin/gh/guilhermeleobas/202/orig 2025-07-17T08:34:50.8765271Z * [new branch] gh/guilhermeleobas/203/base -> origin/gh/guilhermeleobas/203/base 2025-07-17T08:34:50.8768573Z * [new branch] gh/guilhermeleobas/203/head -> origin/gh/guilhermeleobas/203/head 2025-07-17T08:34:50.8771889Z * [new branch] gh/guilhermeleobas/203/orig -> origin/gh/guilhermeleobas/203/orig 2025-07-17T08:34:50.8776667Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-07-17T08:34:50.8779862Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-07-17T08:34:50.8783437Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-07-17T08:34:50.8788535Z * [new branch] gh/guilhermeleobas/205/base -> origin/gh/guilhermeleobas/205/base 2025-07-17T08:34:50.8791899Z * [new branch] gh/guilhermeleobas/205/head -> origin/gh/guilhermeleobas/205/head 2025-07-17T08:34:50.8795464Z * [new branch] gh/guilhermeleobas/205/orig -> origin/gh/guilhermeleobas/205/orig 2025-07-17T08:34:50.8799926Z * [new branch] gh/guilhermeleobas/206/base -> origin/gh/guilhermeleobas/206/base 2025-07-17T08:34:50.8803311Z * [new branch] gh/guilhermeleobas/206/head -> origin/gh/guilhermeleobas/206/head 2025-07-17T08:34:50.8806708Z * [new branch] gh/guilhermeleobas/206/orig -> origin/gh/guilhermeleobas/206/orig 2025-07-17T08:34:50.8811356Z * [new branch] gh/guilhermeleobas/73/base -> origin/gh/guilhermeleobas/73/base 2025-07-17T08:34:50.8814655Z * [new branch] gh/guilhermeleobas/73/head -> origin/gh/guilhermeleobas/73/head 2025-07-17T08:34:50.8817942Z * [new branch] gh/guilhermeleobas/73/orig -> origin/gh/guilhermeleobas/73/orig 2025-07-17T08:34:50.8823836Z * [new branch] gh/henrylhtsang/100/base -> origin/gh/henrylhtsang/100/base 2025-07-17T08:34:50.8827244Z * [new branch] gh/henrylhtsang/100/head -> origin/gh/henrylhtsang/100/head 2025-07-17T08:34:50.8830662Z * [new branch] gh/henrylhtsang/100/orig -> origin/gh/henrylhtsang/100/orig 2025-07-17T08:34:50.8835498Z * [new branch] gh/henrylhtsang/101/base -> origin/gh/henrylhtsang/101/base 2025-07-17T08:34:50.8838888Z * [new branch] gh/henrylhtsang/101/head -> origin/gh/henrylhtsang/101/head 2025-07-17T08:34:50.8842369Z * [new branch] gh/henrylhtsang/101/orig -> origin/gh/henrylhtsang/101/orig 2025-07-17T08:34:50.8846983Z * [new branch] gh/henrylhtsang/102/base -> origin/gh/henrylhtsang/102/base 2025-07-17T08:34:50.8850335Z * [new branch] gh/henrylhtsang/102/head -> origin/gh/henrylhtsang/102/head 2025-07-17T08:34:50.8853657Z * [new branch] gh/henrylhtsang/102/orig -> origin/gh/henrylhtsang/102/orig 2025-07-17T08:34:50.8858308Z * [new branch] gh/henrylhtsang/103/base -> origin/gh/henrylhtsang/103/base 2025-07-17T08:34:50.8861761Z * [new branch] gh/henrylhtsang/103/head -> origin/gh/henrylhtsang/103/head 2025-07-17T08:34:50.8865192Z * [new branch] gh/henrylhtsang/103/orig -> origin/gh/henrylhtsang/103/orig 2025-07-17T08:34:50.8869885Z * [new branch] gh/henrylhtsang/104/base -> origin/gh/henrylhtsang/104/base 2025-07-17T08:34:50.8873269Z * [new branch] gh/henrylhtsang/104/head -> origin/gh/henrylhtsang/104/head 2025-07-17T08:34:50.8876940Z * [new branch] gh/henrylhtsang/104/orig -> origin/gh/henrylhtsang/104/orig 2025-07-17T08:34:50.8881872Z * [new branch] gh/henrylhtsang/105/base -> origin/gh/henrylhtsang/105/base 2025-07-17T08:34:50.8885260Z * [new branch] gh/henrylhtsang/105/head -> origin/gh/henrylhtsang/105/head 2025-07-17T08:34:50.8888646Z * [new branch] gh/henrylhtsang/105/orig -> origin/gh/henrylhtsang/105/orig 2025-07-17T08:34:50.8893228Z * [new branch] gh/henrylhtsang/106/base -> origin/gh/henrylhtsang/106/base 2025-07-17T08:34:50.8896658Z * [new branch] gh/henrylhtsang/106/head -> origin/gh/henrylhtsang/106/head 2025-07-17T08:34:50.8900059Z * [new branch] gh/henrylhtsang/106/orig -> origin/gh/henrylhtsang/106/orig 2025-07-17T08:34:50.8904894Z * [new branch] gh/henrylhtsang/107/base -> origin/gh/henrylhtsang/107/base 2025-07-17T08:34:50.8908345Z * [new branch] gh/henrylhtsang/107/head -> origin/gh/henrylhtsang/107/head 2025-07-17T08:34:50.8911621Z * [new branch] gh/henrylhtsang/107/orig -> origin/gh/henrylhtsang/107/orig 2025-07-17T08:34:50.8916365Z * [new branch] gh/henrylhtsang/108/base -> origin/gh/henrylhtsang/108/base 2025-07-17T08:34:50.8919676Z * [new branch] gh/henrylhtsang/108/head -> origin/gh/henrylhtsang/108/head 2025-07-17T08:34:50.8923081Z * [new branch] gh/henrylhtsang/108/orig -> origin/gh/henrylhtsang/108/orig 2025-07-17T08:34:50.8927678Z * [new branch] gh/henrylhtsang/109/base -> origin/gh/henrylhtsang/109/base 2025-07-17T08:34:50.8931022Z * [new branch] gh/henrylhtsang/109/head -> origin/gh/henrylhtsang/109/head 2025-07-17T08:34:50.8934354Z * [new branch] gh/henrylhtsang/109/orig -> origin/gh/henrylhtsang/109/orig 2025-07-17T08:34:50.8938735Z * [new branch] gh/henrylhtsang/110/base -> origin/gh/henrylhtsang/110/base 2025-07-17T08:34:50.8942124Z * [new branch] gh/henrylhtsang/110/head -> origin/gh/henrylhtsang/110/head 2025-07-17T08:34:50.8945539Z * [new branch] gh/henrylhtsang/110/orig -> origin/gh/henrylhtsang/110/orig 2025-07-17T08:34:50.8949951Z * [new branch] gh/henrylhtsang/111/base -> origin/gh/henrylhtsang/111/base 2025-07-17T08:34:50.8953268Z * [new branch] gh/henrylhtsang/111/head -> origin/gh/henrylhtsang/111/head 2025-07-17T08:34:50.8956853Z * [new branch] gh/henrylhtsang/111/orig -> origin/gh/henrylhtsang/111/orig 2025-07-17T08:34:50.8961761Z * [new branch] gh/henrylhtsang/112/base -> origin/gh/henrylhtsang/112/base 2025-07-17T08:34:50.8965106Z * [new branch] gh/henrylhtsang/112/head -> origin/gh/henrylhtsang/112/head 2025-07-17T08:34:50.8968524Z * [new branch] gh/henrylhtsang/112/orig -> origin/gh/henrylhtsang/112/orig 2025-07-17T08:34:50.8973185Z * [new branch] gh/henrylhtsang/113/base -> origin/gh/henrylhtsang/113/base 2025-07-17T08:34:50.8976654Z * [new branch] gh/henrylhtsang/113/head -> origin/gh/henrylhtsang/113/head 2025-07-17T08:34:50.8980019Z * [new branch] gh/henrylhtsang/113/orig -> origin/gh/henrylhtsang/113/orig 2025-07-17T08:34:50.8984636Z * [new branch] gh/henrylhtsang/114/base -> origin/gh/henrylhtsang/114/base 2025-07-17T08:34:50.8988015Z * [new branch] gh/henrylhtsang/114/head -> origin/gh/henrylhtsang/114/head 2025-07-17T08:34:50.8991349Z * [new branch] gh/henrylhtsang/114/orig -> origin/gh/henrylhtsang/114/orig 2025-07-17T08:34:50.8996259Z * [new branch] gh/henrylhtsang/115/base -> origin/gh/henrylhtsang/115/base 2025-07-17T08:34:50.8999607Z * [new branch] gh/henrylhtsang/115/head -> origin/gh/henrylhtsang/115/head 2025-07-17T08:34:50.9002956Z * [new branch] gh/henrylhtsang/115/orig -> origin/gh/henrylhtsang/115/orig 2025-07-17T08:34:50.9007762Z * [new branch] gh/henrylhtsang/116/base -> origin/gh/henrylhtsang/116/base 2025-07-17T08:34:50.9011168Z * [new branch] gh/henrylhtsang/116/head -> origin/gh/henrylhtsang/116/head 2025-07-17T08:34:50.9014497Z * [new branch] gh/henrylhtsang/116/orig -> origin/gh/henrylhtsang/116/orig 2025-07-17T08:34:50.9019094Z * [new branch] gh/henrylhtsang/117/base -> origin/gh/henrylhtsang/117/base 2025-07-17T08:34:50.9022436Z * [new branch] gh/henrylhtsang/117/head -> origin/gh/henrylhtsang/117/head 2025-07-17T08:34:50.9025848Z * [new branch] gh/henrylhtsang/117/orig -> origin/gh/henrylhtsang/117/orig 2025-07-17T08:34:50.9030405Z * [new branch] gh/henrylhtsang/118/base -> origin/gh/henrylhtsang/118/base 2025-07-17T08:34:50.9033727Z * [new branch] gh/henrylhtsang/118/head -> origin/gh/henrylhtsang/118/head 2025-07-17T08:34:50.9037495Z * [new branch] gh/henrylhtsang/118/orig -> origin/gh/henrylhtsang/118/orig 2025-07-17T08:34:50.9041851Z * [new branch] gh/henrylhtsang/119/base -> origin/gh/henrylhtsang/119/base 2025-07-17T08:34:50.9045222Z * [new branch] gh/henrylhtsang/119/head -> origin/gh/henrylhtsang/119/head 2025-07-17T08:34:50.9048563Z * [new branch] gh/henrylhtsang/119/orig -> origin/gh/henrylhtsang/119/orig 2025-07-17T08:34:50.9053123Z * [new branch] gh/henrylhtsang/120/base -> origin/gh/henrylhtsang/120/base 2025-07-17T08:34:50.9056517Z * [new branch] gh/henrylhtsang/120/head -> origin/gh/henrylhtsang/120/head 2025-07-17T08:34:50.9059868Z * [new branch] gh/henrylhtsang/120/orig -> origin/gh/henrylhtsang/120/orig 2025-07-17T08:34:50.9064598Z * [new branch] gh/henrylhtsang/121/base -> origin/gh/henrylhtsang/121/base 2025-07-17T08:34:50.9067996Z * [new branch] gh/henrylhtsang/121/head -> origin/gh/henrylhtsang/121/head 2025-07-17T08:34:50.9071646Z * [new branch] gh/henrylhtsang/121/orig -> origin/gh/henrylhtsang/121/orig 2025-07-17T08:34:50.9076472Z * [new branch] gh/henrylhtsang/122/base -> origin/gh/henrylhtsang/122/base 2025-07-17T08:34:50.9079840Z * [new branch] gh/henrylhtsang/122/head -> origin/gh/henrylhtsang/122/head 2025-07-17T08:34:50.9083263Z * [new branch] gh/henrylhtsang/122/orig -> origin/gh/henrylhtsang/122/orig 2025-07-17T08:34:50.9088032Z * [new branch] gh/henrylhtsang/123/base -> origin/gh/henrylhtsang/123/base 2025-07-17T08:34:50.9091553Z * [new branch] gh/henrylhtsang/123/head -> origin/gh/henrylhtsang/123/head 2025-07-17T08:34:50.9094914Z * [new branch] gh/henrylhtsang/123/orig -> origin/gh/henrylhtsang/123/orig 2025-07-17T08:34:50.9099729Z * [new branch] gh/henrylhtsang/97/base -> origin/gh/henrylhtsang/97/base 2025-07-17T08:34:50.9103167Z * [new branch] gh/henrylhtsang/97/head -> origin/gh/henrylhtsang/97/head 2025-07-17T08:34:50.9106592Z * [new branch] gh/henrylhtsang/97/orig -> origin/gh/henrylhtsang/97/orig 2025-07-17T08:34:50.9111285Z * [new branch] gh/henrylhtsang/98/base -> origin/gh/henrylhtsang/98/base 2025-07-17T08:34:50.9114857Z * [new branch] gh/henrylhtsang/98/head -> origin/gh/henrylhtsang/98/head 2025-07-17T08:34:50.9118257Z * [new branch] gh/henrylhtsang/98/orig -> origin/gh/henrylhtsang/98/orig 2025-07-17T08:34:50.9122810Z * [new branch] gh/henrylhtsang/99/base -> origin/gh/henrylhtsang/99/base 2025-07-17T08:34:50.9126201Z * [new branch] gh/henrylhtsang/99/head -> origin/gh/henrylhtsang/99/head 2025-07-17T08:34:50.9129585Z * [new branch] gh/henrylhtsang/99/orig -> origin/gh/henrylhtsang/99/orig 2025-07-17T08:34:50.9135149Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-07-17T08:34:50.9138550Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-07-17T08:34:50.9144114Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-07-17T08:34:50.9147437Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-07-17T08:34:50.9151967Z * [new branch] gh/isuruf/105/base -> origin/gh/isuruf/105/base 2025-07-17T08:34:50.9155459Z * [new branch] gh/isuruf/105/head -> origin/gh/isuruf/105/head 2025-07-17T08:34:50.9158818Z * [new branch] gh/isuruf/105/orig -> origin/gh/isuruf/105/orig 2025-07-17T08:34:50.9163371Z * [new branch] gh/isuruf/110/base -> origin/gh/isuruf/110/base 2025-07-17T08:34:50.9166688Z * [new branch] gh/isuruf/110/head -> origin/gh/isuruf/110/head 2025-07-17T08:34:50.9170196Z * [new branch] gh/isuruf/110/orig -> origin/gh/isuruf/110/orig 2025-07-17T08:34:50.9174556Z * [new branch] gh/isuruf/116/base -> origin/gh/isuruf/116/base 2025-07-17T08:34:50.9177870Z * [new branch] gh/isuruf/116/head -> origin/gh/isuruf/116/head 2025-07-17T08:34:50.9181268Z * [new branch] gh/isuruf/116/orig -> origin/gh/isuruf/116/orig 2025-07-17T08:34:50.9186009Z * [new branch] gh/isuruf/141/base -> origin/gh/isuruf/141/base 2025-07-17T08:34:50.9189377Z * [new branch] gh/isuruf/141/head -> origin/gh/isuruf/141/head 2025-07-17T08:34:50.9192714Z * [new branch] gh/isuruf/141/orig -> origin/gh/isuruf/141/orig 2025-07-17T08:34:50.9197465Z * [new branch] gh/isuruf/143/base -> origin/gh/isuruf/143/base 2025-07-17T08:34:50.9200853Z * [new branch] gh/isuruf/143/head -> origin/gh/isuruf/143/head 2025-07-17T08:34:50.9204219Z * [new branch] gh/isuruf/143/orig -> origin/gh/isuruf/143/orig 2025-07-17T08:34:50.9208958Z * [new branch] gh/isuruf/145/base -> origin/gh/isuruf/145/base 2025-07-17T08:34:50.9212175Z * [new branch] gh/isuruf/145/head -> origin/gh/isuruf/145/head 2025-07-17T08:34:50.9215485Z * [new branch] gh/isuruf/145/orig -> origin/gh/isuruf/145/orig 2025-07-17T08:34:50.9220310Z * [new branch] gh/isuruf/147/base -> origin/gh/isuruf/147/base 2025-07-17T08:34:50.9223619Z * [new branch] gh/isuruf/147/head -> origin/gh/isuruf/147/head 2025-07-17T08:34:50.9226934Z * [new branch] gh/isuruf/147/orig -> origin/gh/isuruf/147/orig 2025-07-17T08:34:50.9231603Z * [new branch] gh/isuruf/148/base -> origin/gh/isuruf/148/base 2025-07-17T08:34:50.9235038Z * [new branch] gh/isuruf/148/head -> origin/gh/isuruf/148/head 2025-07-17T08:34:50.9238481Z * [new branch] gh/isuruf/148/orig -> origin/gh/isuruf/148/orig 2025-07-17T08:34:50.9243901Z * [new branch] gh/jamesjwu/140/base -> origin/gh/jamesjwu/140/base 2025-07-17T08:34:50.9247233Z * [new branch] gh/jamesjwu/140/head -> origin/gh/jamesjwu/140/head 2025-07-17T08:34:50.9250617Z * [new branch] gh/jamesjwu/140/orig -> origin/gh/jamesjwu/140/orig 2025-07-17T08:34:50.9255189Z * [new branch] gh/jamesjwu/150/base -> origin/gh/jamesjwu/150/base 2025-07-17T08:34:50.9258533Z * [new branch] gh/jamesjwu/150/head -> origin/gh/jamesjwu/150/head 2025-07-17T08:34:50.9261857Z * [new branch] gh/jamesjwu/150/orig -> origin/gh/jamesjwu/150/orig 2025-07-17T08:34:50.9266771Z * [new branch] gh/jamesjwu/154/base -> origin/gh/jamesjwu/154/base 2025-07-17T08:34:50.9270049Z * [new branch] gh/jamesjwu/154/head -> origin/gh/jamesjwu/154/head 2025-07-17T08:34:50.9273491Z * [new branch] gh/jamesjwu/154/orig -> origin/gh/jamesjwu/154/orig 2025-07-17T08:34:50.9278184Z * [new branch] gh/jamesjwu/155/base -> origin/gh/jamesjwu/155/base 2025-07-17T08:34:50.9281566Z * [new branch] gh/jamesjwu/155/head -> origin/gh/jamesjwu/155/head 2025-07-17T08:34:50.9284875Z * [new branch] gh/jamesjwu/155/orig -> origin/gh/jamesjwu/155/orig 2025-07-17T08:34:50.9289590Z * [new branch] gh/jamesjwu/159/base -> origin/gh/jamesjwu/159/base 2025-07-17T08:34:50.9292921Z * [new branch] gh/jamesjwu/159/head -> origin/gh/jamesjwu/159/head 2025-07-17T08:34:50.9296271Z * [new branch] gh/jamesjwu/159/orig -> origin/gh/jamesjwu/159/orig 2025-07-17T08:34:50.9301152Z * [new branch] gh/jamesjwu/163/base -> origin/gh/jamesjwu/163/base 2025-07-17T08:34:50.9304758Z * [new branch] gh/jamesjwu/163/head -> origin/gh/jamesjwu/163/head 2025-07-17T08:34:50.9307856Z * [new branch] gh/jamesjwu/163/orig -> origin/gh/jamesjwu/163/orig 2025-07-17T08:34:50.9312594Z * [new branch] gh/jamesjwu/164/base -> origin/gh/jamesjwu/164/base 2025-07-17T08:34:50.9316079Z * [new branch] gh/jamesjwu/164/head -> origin/gh/jamesjwu/164/head 2025-07-17T08:34:50.9319533Z * [new branch] gh/jamesjwu/164/orig -> origin/gh/jamesjwu/164/orig 2025-07-17T08:34:50.9324246Z * [new branch] gh/jamesjwu/165/base -> origin/gh/jamesjwu/165/base 2025-07-17T08:34:50.9327503Z * [new branch] gh/jamesjwu/165/head -> origin/gh/jamesjwu/165/head 2025-07-17T08:34:50.9330867Z * [new branch] gh/jamesjwu/165/orig -> origin/gh/jamesjwu/165/orig 2025-07-17T08:34:50.9335517Z * [new branch] gh/jamesjwu/166/base -> origin/gh/jamesjwu/166/base 2025-07-17T08:34:50.9339029Z * [new branch] gh/jamesjwu/166/head -> origin/gh/jamesjwu/166/head 2025-07-17T08:34:50.9342277Z * [new branch] gh/jamesjwu/166/orig -> origin/gh/jamesjwu/166/orig 2025-07-17T08:34:50.9346950Z * [new branch] gh/jamesjwu/167/base -> origin/gh/jamesjwu/167/base 2025-07-17T08:34:50.9350592Z * [new branch] gh/jamesjwu/167/head -> origin/gh/jamesjwu/167/head 2025-07-17T08:34:50.9354118Z * [new branch] gh/jamesjwu/167/orig -> origin/gh/jamesjwu/167/orig 2025-07-17T08:34:50.9358880Z * [new branch] gh/jamesjwu/168/base -> origin/gh/jamesjwu/168/base 2025-07-17T08:34:50.9362190Z * [new branch] gh/jamesjwu/168/head -> origin/gh/jamesjwu/168/head 2025-07-17T08:34:50.9365634Z * [new branch] gh/jamesjwu/168/orig -> origin/gh/jamesjwu/168/orig 2025-07-17T08:34:50.9370240Z * [new branch] gh/jamesjwu/169/base -> origin/gh/jamesjwu/169/base 2025-07-17T08:34:50.9373631Z * [new branch] gh/jamesjwu/169/head -> origin/gh/jamesjwu/169/head 2025-07-17T08:34:50.9376993Z * [new branch] gh/jamesjwu/169/orig -> origin/gh/jamesjwu/169/orig 2025-07-17T08:34:50.9381552Z * [new branch] gh/jamesjwu/170/base -> origin/gh/jamesjwu/170/base 2025-07-17T08:34:50.9384894Z * [new branch] gh/jamesjwu/170/head -> origin/gh/jamesjwu/170/head 2025-07-17T08:34:50.9388353Z * [new branch] gh/jamesjwu/170/orig -> origin/gh/jamesjwu/170/orig 2025-07-17T08:34:50.9392863Z * [new branch] gh/jamesjwu/171/base -> origin/gh/jamesjwu/171/base 2025-07-17T08:34:50.9396558Z * [new branch] gh/jamesjwu/171/head -> origin/gh/jamesjwu/171/head 2025-07-17T08:34:50.9399865Z * [new branch] gh/jamesjwu/171/orig -> origin/gh/jamesjwu/171/orig 2025-07-17T08:34:50.9404561Z * [new branch] gh/jamesjwu/172/base -> origin/gh/jamesjwu/172/base 2025-07-17T08:34:50.9407893Z * [new branch] gh/jamesjwu/172/head -> origin/gh/jamesjwu/172/head 2025-07-17T08:34:50.9411322Z * [new branch] gh/jamesjwu/172/orig -> origin/gh/jamesjwu/172/orig 2025-07-17T08:34:50.9415930Z * [new branch] gh/jamesjwu/173/base -> origin/gh/jamesjwu/173/base 2025-07-17T08:34:50.9419367Z * [new branch] gh/jamesjwu/173/head -> origin/gh/jamesjwu/173/head 2025-07-17T08:34:50.9422725Z * [new branch] gh/jamesjwu/173/orig -> origin/gh/jamesjwu/173/orig 2025-07-17T08:34:50.9427273Z * [new branch] gh/jamesjwu/174/base -> origin/gh/jamesjwu/174/base 2025-07-17T08:34:50.9430600Z * [new branch] gh/jamesjwu/174/head -> origin/gh/jamesjwu/174/head 2025-07-17T08:34:50.9434035Z * [new branch] gh/jamesjwu/174/orig -> origin/gh/jamesjwu/174/orig 2025-07-17T08:34:50.9438961Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-07-17T08:34:50.9442243Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-07-17T08:34:50.9446633Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-07-17T08:34:50.9449927Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-07-17T08:34:50.9454421Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-07-17T08:34:50.9458017Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-07-17T08:34:50.9462478Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-07-17T08:34:50.9465833Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-07-17T08:34:50.9470244Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-07-17T08:34:50.9473583Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-07-17T08:34:50.9478174Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-07-17T08:34:50.9481483Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-07-17T08:34:50.9486033Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-07-17T08:34:50.9489382Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-07-17T08:34:50.9493883Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-07-17T08:34:50.9497343Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-07-17T08:34:50.9501755Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-07-17T08:34:50.9505032Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-07-17T08:34:50.9509437Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-07-17T08:34:50.9512745Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-07-17T08:34:50.9517359Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-07-17T08:34:50.9520670Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-07-17T08:34:50.9525091Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-07-17T08:34:50.9528483Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-07-17T08:34:50.9533085Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-07-17T08:34:50.9536450Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-07-17T08:34:50.9540943Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-07-17T08:34:50.9544236Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-07-17T08:34:50.9550012Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-07-17T08:34:50.9553343Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-07-17T08:34:50.9556952Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-07-17T08:34:50.9561720Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-07-17T08:34:50.9564804Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-07-17T08:34:50.9568323Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-07-17T08:34:50.9572924Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-07-17T08:34:50.9576362Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-07-17T08:34:50.9579938Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-07-17T08:34:50.9584403Z * [new branch] gh/janeyx99/256/base -> origin/gh/janeyx99/256/base 2025-07-17T08:34:50.9587771Z * [new branch] gh/janeyx99/256/head -> origin/gh/janeyx99/256/head 2025-07-17T08:34:50.9591247Z * [new branch] gh/janeyx99/256/orig -> origin/gh/janeyx99/256/orig 2025-07-17T08:34:50.9596252Z * [new branch] gh/janeyx99/257/base -> origin/gh/janeyx99/257/base 2025-07-17T08:34:50.9603073Z * [new branch] gh/janeyx99/257/head -> origin/gh/janeyx99/257/head 2025-07-17T08:34:50.9603381Z * [new branch] gh/janeyx99/257/orig -> origin/gh/janeyx99/257/orig 2025-07-17T08:34:50.9607592Z * [new branch] gh/janeyx99/259/base -> origin/gh/janeyx99/259/base 2025-07-17T08:34:50.9610857Z * [new branch] gh/janeyx99/259/head -> origin/gh/janeyx99/259/head 2025-07-17T08:34:50.9614249Z * [new branch] gh/janeyx99/259/orig -> origin/gh/janeyx99/259/orig 2025-07-17T08:34:50.9618826Z * [new branch] gh/janeyx99/260/base -> origin/gh/janeyx99/260/base 2025-07-17T08:34:50.9622182Z * [new branch] gh/janeyx99/260/head -> origin/gh/janeyx99/260/head 2025-07-17T08:34:50.9625515Z * [new branch] gh/janeyx99/260/orig -> origin/gh/janeyx99/260/orig 2025-07-17T08:34:50.9630153Z * [new branch] gh/janeyx99/261/base -> origin/gh/janeyx99/261/base 2025-07-17T08:34:50.9633702Z * [new branch] gh/janeyx99/261/head -> origin/gh/janeyx99/261/head 2025-07-17T08:34:50.9638509Z * [new branch] gh/janeyx99/261/orig -> origin/gh/janeyx99/261/orig 2025-07-17T08:34:50.9643041Z * [new branch] gh/janeyx99/262/base -> origin/gh/janeyx99/262/base 2025-07-17T08:34:50.9646426Z * [new branch] gh/janeyx99/262/head -> origin/gh/janeyx99/262/head 2025-07-17T08:34:50.9650871Z * [new branch] gh/janeyx99/263/base -> origin/gh/janeyx99/263/base 2025-07-17T08:34:50.9654360Z * [new branch] gh/janeyx99/263/head -> origin/gh/janeyx99/263/head 2025-07-17T08:34:50.9658766Z * [new branch] gh/janeyx99/264/base -> origin/gh/janeyx99/264/base 2025-07-17T08:34:50.9662058Z * [new branch] gh/janeyx99/264/head -> origin/gh/janeyx99/264/head 2025-07-17T08:34:50.9666489Z * [new branch] gh/janeyx99/265/base -> origin/gh/janeyx99/265/base 2025-07-17T08:34:50.9669784Z * [new branch] gh/janeyx99/265/head -> origin/gh/janeyx99/265/head 2025-07-17T08:34:50.9681013Z * [new branch] gh/janeyx99/266/base -> origin/gh/janeyx99/266/base 2025-07-17T08:34:50.9684491Z * [new branch] gh/janeyx99/266/head -> origin/gh/janeyx99/266/head 2025-07-17T08:34:50.9687854Z * [new branch] gh/janeyx99/266/orig -> origin/gh/janeyx99/266/orig 2025-07-17T08:34:50.9692614Z * [new branch] gh/janeyx99/267/base -> origin/gh/janeyx99/267/base 2025-07-17T08:34:50.9695971Z * [new branch] gh/janeyx99/267/head -> origin/gh/janeyx99/267/head 2025-07-17T08:34:50.9699347Z * [new branch] gh/janeyx99/267/orig -> origin/gh/janeyx99/267/orig 2025-07-17T08:34:50.9703943Z * [new branch] gh/janeyx99/268/base -> origin/gh/janeyx99/268/base 2025-07-17T08:34:50.9707270Z * [new branch] gh/janeyx99/268/head -> origin/gh/janeyx99/268/head 2025-07-17T08:34:50.9710601Z * [new branch] gh/janeyx99/268/orig -> origin/gh/janeyx99/268/orig 2025-07-17T08:34:50.9715378Z * [new branch] gh/janeyx99/269/base -> origin/gh/janeyx99/269/base 2025-07-17T08:34:50.9718767Z * [new branch] gh/janeyx99/269/head -> origin/gh/janeyx99/269/head 2025-07-17T08:34:50.9722305Z * [new branch] gh/janeyx99/269/orig -> origin/gh/janeyx99/269/orig 2025-07-17T08:34:50.9726837Z * [new branch] gh/janeyx99/270/base -> origin/gh/janeyx99/270/base 2025-07-17T08:34:50.9730182Z * [new branch] gh/janeyx99/270/head -> origin/gh/janeyx99/270/head 2025-07-17T08:34:50.9733593Z * [new branch] gh/janeyx99/270/orig -> origin/gh/janeyx99/270/orig 2025-07-17T08:34:50.9738202Z * [new branch] gh/janeyx99/271/base -> origin/gh/janeyx99/271/base 2025-07-17T08:34:50.9743110Z * [new branch] gh/janeyx99/271/head -> origin/gh/janeyx99/271/head 2025-07-17T08:34:50.9744891Z * [new branch] gh/janeyx99/271/orig -> origin/gh/janeyx99/271/orig 2025-07-17T08:34:50.9749610Z * [new branch] gh/janeyx99/272/base -> origin/gh/janeyx99/272/base 2025-07-17T08:34:50.9752950Z * [new branch] gh/janeyx99/272/head -> origin/gh/janeyx99/272/head 2025-07-17T08:34:50.9756395Z * [new branch] gh/janeyx99/272/orig -> origin/gh/janeyx99/272/orig 2025-07-17T08:34:50.9761032Z * [new branch] gh/janeyx99/273/base -> origin/gh/janeyx99/273/base 2025-07-17T08:34:50.9764376Z * [new branch] gh/janeyx99/273/head -> origin/gh/janeyx99/273/head 2025-07-17T08:34:50.9767736Z * [new branch] gh/janeyx99/273/orig -> origin/gh/janeyx99/273/orig 2025-07-17T08:34:50.9772428Z * [new branch] gh/janeyx99/274/base -> origin/gh/janeyx99/274/base 2025-07-17T08:34:50.9775749Z * [new branch] gh/janeyx99/274/head -> origin/gh/janeyx99/274/head 2025-07-17T08:34:50.9779177Z * [new branch] gh/janeyx99/274/orig -> origin/gh/janeyx99/274/orig 2025-07-17T08:34:50.9783598Z * [new branch] gh/janeyx99/275/base -> origin/gh/janeyx99/275/base 2025-07-17T08:34:50.9786938Z * [new branch] gh/janeyx99/275/head -> origin/gh/janeyx99/275/head 2025-07-17T08:34:50.9790292Z * [new branch] gh/janeyx99/275/orig -> origin/gh/janeyx99/275/orig 2025-07-17T08:34:50.9794982Z * [new branch] gh/janeyx99/276/base -> origin/gh/janeyx99/276/base 2025-07-17T08:34:50.9798405Z * [new branch] gh/janeyx99/276/head -> origin/gh/janeyx99/276/head 2025-07-17T08:34:50.9801756Z * [new branch] gh/janeyx99/276/orig -> origin/gh/janeyx99/276/orig 2025-07-17T08:34:50.9806327Z * [new branch] gh/janeyx99/277/base -> origin/gh/janeyx99/277/base 2025-07-17T08:34:50.9809745Z * [new branch] gh/janeyx99/277/head -> origin/gh/janeyx99/277/head 2025-07-17T08:34:50.9813154Z * [new branch] gh/janeyx99/277/orig -> origin/gh/janeyx99/277/orig 2025-07-17T08:34:50.9817852Z * [new branch] gh/janeyx99/278/base -> origin/gh/janeyx99/278/base 2025-07-17T08:34:50.9821176Z * [new branch] gh/janeyx99/278/head -> origin/gh/janeyx99/278/head 2025-07-17T08:34:50.9824565Z * [new branch] gh/janeyx99/278/orig -> origin/gh/janeyx99/278/orig 2025-07-17T08:34:50.9829163Z * [new branch] gh/janeyx99/279/base -> origin/gh/janeyx99/279/base 2025-07-17T08:34:50.9832414Z * [new branch] gh/janeyx99/279/head -> origin/gh/janeyx99/279/head 2025-07-17T08:34:50.9835967Z * [new branch] gh/janeyx99/279/orig -> origin/gh/janeyx99/279/orig 2025-07-17T08:34:50.9840519Z * [new branch] gh/janeyx99/280/base -> origin/gh/janeyx99/280/base 2025-07-17T08:34:50.9843853Z * [new branch] gh/janeyx99/280/head -> origin/gh/janeyx99/280/head 2025-07-17T08:34:50.9847201Z * [new branch] gh/janeyx99/280/orig -> origin/gh/janeyx99/280/orig 2025-07-17T08:34:50.9852090Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-07-17T08:34:50.9855606Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-07-17T08:34:50.9858806Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-07-17T08:34:50.9864510Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-07-17T08:34:50.9867818Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-07-17T08:34:50.9872373Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-07-17T08:34:50.9876191Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-07-17T08:34:50.9879604Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-07-17T08:34:50.9884164Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-07-17T08:34:50.9887560Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-07-17T08:34:50.9890930Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-07-17T08:34:50.9895482Z * [new branch] gh/jansel/513/base -> origin/gh/jansel/513/base 2025-07-17T08:34:50.9898757Z * [new branch] gh/jansel/513/head -> origin/gh/jansel/513/head 2025-07-17T08:34:50.9902171Z * [new branch] gh/jansel/513/orig -> origin/gh/jansel/513/orig 2025-07-17T08:34:50.9906872Z * [new branch] gh/jansel/519/base -> origin/gh/jansel/519/base 2025-07-17T08:34:50.9910234Z * [new branch] gh/jansel/519/head -> origin/gh/jansel/519/head 2025-07-17T08:34:50.9913563Z * [new branch] gh/jansel/519/orig -> origin/gh/jansel/519/orig 2025-07-17T08:34:50.9918337Z * [new branch] gh/jansel/520/base -> origin/gh/jansel/520/base 2025-07-17T08:34:50.9921669Z * [new branch] gh/jansel/520/head -> origin/gh/jansel/520/head 2025-07-17T08:34:50.9925047Z * [new branch] gh/jansel/520/orig -> origin/gh/jansel/520/orig 2025-07-17T08:34:50.9929554Z * [new branch] gh/jansel/521/base -> origin/gh/jansel/521/base 2025-07-17T08:34:50.9932881Z * [new branch] gh/jansel/521/head -> origin/gh/jansel/521/head 2025-07-17T08:34:50.9936221Z * [new branch] gh/jansel/521/orig -> origin/gh/jansel/521/orig 2025-07-17T08:34:50.9940841Z * [new branch] gh/jansel/522/base -> origin/gh/jansel/522/base 2025-07-17T08:34:50.9944174Z * [new branch] gh/jansel/522/head -> origin/gh/jansel/522/head 2025-07-17T08:34:50.9947482Z * [new branch] gh/jansel/522/orig -> origin/gh/jansel/522/orig 2025-07-17T08:34:50.9952158Z * [new branch] gh/jansel/523/base -> origin/gh/jansel/523/base 2025-07-17T08:34:50.9956034Z * [new branch] gh/jansel/523/head -> origin/gh/jansel/523/head 2025-07-17T08:34:50.9959329Z * [new branch] gh/jansel/523/orig -> origin/gh/jansel/523/orig 2025-07-17T08:34:50.9963860Z * [new branch] gh/jansel/524/base -> origin/gh/jansel/524/base 2025-07-17T08:34:50.9967233Z * [new branch] gh/jansel/524/head -> origin/gh/jansel/524/head 2025-07-17T08:34:50.9970618Z * [new branch] gh/jansel/524/orig -> origin/gh/jansel/524/orig 2025-07-17T08:34:50.9975256Z * [new branch] gh/jansel/525/base -> origin/gh/jansel/525/base 2025-07-17T08:34:50.9978618Z * [new branch] gh/jansel/525/head -> origin/gh/jansel/525/head 2025-07-17T08:34:50.9981947Z * [new branch] gh/jansel/525/orig -> origin/gh/jansel/525/orig 2025-07-17T08:34:50.9986473Z * [new branch] gh/jansel/526/base -> origin/gh/jansel/526/base 2025-07-17T08:34:50.9990112Z * [new branch] gh/jansel/526/head -> origin/gh/jansel/526/head 2025-07-17T08:34:50.9993273Z * [new branch] gh/jansel/526/orig -> origin/gh/jansel/526/orig 2025-07-17T08:34:50.9998051Z * [new branch] gh/jansel/527/base -> origin/gh/jansel/527/base 2025-07-17T08:34:51.0001394Z * [new branch] gh/jansel/527/head -> origin/gh/jansel/527/head 2025-07-17T08:34:51.0004728Z * [new branch] gh/jansel/527/orig -> origin/gh/jansel/527/orig 2025-07-17T08:34:51.0009251Z * [new branch] gh/jansel/528/base -> origin/gh/jansel/528/base 2025-07-17T08:34:51.0012576Z * [new branch] gh/jansel/528/head -> origin/gh/jansel/528/head 2025-07-17T08:34:51.0015897Z * [new branch] gh/jansel/528/orig -> origin/gh/jansel/528/orig 2025-07-17T08:34:51.0020403Z * [new branch] gh/jansel/529/base -> origin/gh/jansel/529/base 2025-07-17T08:34:51.0023717Z * [new branch] gh/jansel/529/head -> origin/gh/jansel/529/head 2025-07-17T08:34:51.0027091Z * [new branch] gh/jansel/529/orig -> origin/gh/jansel/529/orig 2025-07-17T08:34:51.0031722Z * [new branch] gh/jansel/530/base -> origin/gh/jansel/530/base 2025-07-17T08:34:51.0035201Z * [new branch] gh/jansel/530/head -> origin/gh/jansel/530/head 2025-07-17T08:34:51.0038682Z * [new branch] gh/jansel/530/orig -> origin/gh/jansel/530/orig 2025-07-17T08:34:51.0043695Z * [new branch] gh/jansel/531/base -> origin/gh/jansel/531/base 2025-07-17T08:34:51.0047081Z * [new branch] gh/jansel/531/head -> origin/gh/jansel/531/head 2025-07-17T08:34:51.0050442Z * [new branch] gh/jansel/531/orig -> origin/gh/jansel/531/orig 2025-07-17T08:34:51.0054961Z * [new branch] gh/jansel/532/base -> origin/gh/jansel/532/base 2025-07-17T08:34:51.0058306Z * [new branch] gh/jansel/532/head -> origin/gh/jansel/532/head 2025-07-17T08:34:51.0061650Z * [new branch] gh/jansel/532/orig -> origin/gh/jansel/532/orig 2025-07-17T08:34:51.0066166Z * [new branch] gh/jansel/533/base -> origin/gh/jansel/533/base 2025-07-17T08:34:51.0069488Z * [new branch] gh/jansel/533/head -> origin/gh/jansel/533/head 2025-07-17T08:34:51.0072833Z * [new branch] gh/jansel/533/orig -> origin/gh/jansel/533/orig 2025-07-17T08:34:51.0078597Z * [new branch] gh/jbschlosser/226/base -> origin/gh/jbschlosser/226/base 2025-07-17T08:34:51.0081907Z * [new branch] gh/jbschlosser/226/head -> origin/gh/jbschlosser/226/head 2025-07-17T08:34:51.0085476Z * [new branch] gh/jbschlosser/226/orig -> origin/gh/jbschlosser/226/orig 2025-07-17T08:34:51.0090153Z * [new branch] gh/jbschlosser/239/base -> origin/gh/jbschlosser/239/base 2025-07-17T08:34:51.0093519Z * [new branch] gh/jbschlosser/239/head -> origin/gh/jbschlosser/239/head 2025-07-17T08:34:51.0096870Z * [new branch] gh/jbschlosser/239/orig -> origin/gh/jbschlosser/239/orig 2025-07-17T08:34:51.0101427Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-07-17T08:34:51.0104826Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-07-17T08:34:51.0108153Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-07-17T08:34:51.0112742Z * [new branch] gh/jbschlosser/248/base -> origin/gh/jbschlosser/248/base 2025-07-17T08:34:51.0116622Z * [new branch] gh/jbschlosser/248/head -> origin/gh/jbschlosser/248/head 2025-07-17T08:34:51.0120000Z * [new branch] gh/jbschlosser/248/orig -> origin/gh/jbschlosser/248/orig 2025-07-17T08:34:51.0125698Z * [new branch] gh/jerryzh168/1/base -> origin/gh/jerryzh168/1/base 2025-07-17T08:34:51.0129111Z * [new branch] gh/jerryzh168/1/head -> origin/gh/jerryzh168/1/head 2025-07-17T08:34:51.0132430Z * [new branch] gh/jerryzh168/1/orig -> origin/gh/jerryzh168/1/orig 2025-07-17T08:34:51.0138057Z * [new branch] gh/jiayisunx/57/base -> origin/gh/jiayisunx/57/base 2025-07-17T08:34:51.0141385Z * [new branch] gh/jiayisunx/57/head -> origin/gh/jiayisunx/57/head 2025-07-17T08:34:51.0144798Z * [new branch] gh/jiayisunx/57/orig -> origin/gh/jiayisunx/57/orig 2025-07-17T08:34:51.0149397Z * [new branch] gh/jiayisunx/58/base -> origin/gh/jiayisunx/58/base 2025-07-17T08:34:51.0152714Z * [new branch] gh/jiayisunx/58/head -> origin/gh/jiayisunx/58/head 2025-07-17T08:34:51.0156478Z * [new branch] gh/jiayisunx/58/orig -> origin/gh/jiayisunx/58/orig 2025-07-17T08:34:51.0160725Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-07-17T08:34:51.0164095Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-07-17T08:34:51.0167527Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-07-17T08:34:51.0172125Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-07-17T08:34:51.0175536Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-07-17T08:34:51.0179201Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-07-17T08:34:51.0183511Z * [new branch] gh/jiayisunx/63/base -> origin/gh/jiayisunx/63/base 2025-07-17T08:34:51.0186822Z * [new branch] gh/jiayisunx/63/head -> origin/gh/jiayisunx/63/head 2025-07-17T08:34:51.0190355Z * [new branch] gh/jiayisunx/63/orig -> origin/gh/jiayisunx/63/orig 2025-07-17T08:34:51.0196904Z * [new branch] gh/jiayisunx/64/base -> origin/gh/jiayisunx/64/base 2025-07-17T08:34:51.0200033Z * [new branch] gh/jiayisunx/64/head -> origin/gh/jiayisunx/64/head 2025-07-17T08:34:51.0203355Z * [new branch] gh/jiayisunx/64/orig -> origin/gh/jiayisunx/64/orig 2025-07-17T08:34:51.0207917Z * [new branch] gh/jiayisunx/65/base -> origin/gh/jiayisunx/65/base 2025-07-17T08:34:51.0211253Z * [new branch] gh/jiayisunx/65/head -> origin/gh/jiayisunx/65/head 2025-07-17T08:34:51.0214611Z * [new branch] gh/jiayisunx/65/orig -> origin/gh/jiayisunx/65/orig 2025-07-17T08:34:51.0220032Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-07-17T08:34:51.0223328Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-07-17T08:34:51.0228795Z * [new branch] gh/justinchuby/111/base -> origin/gh/justinchuby/111/base 2025-07-17T08:34:51.0232133Z * [new branch] gh/justinchuby/111/head -> origin/gh/justinchuby/111/head 2025-07-17T08:34:51.0236002Z * [new branch] gh/justinchuby/111/orig -> origin/gh/justinchuby/111/orig 2025-07-17T08:34:51.0241664Z * [new branch] gh/kurtamohler/31/base -> origin/gh/kurtamohler/31/base 2025-07-17T08:34:51.0244771Z * [new branch] gh/kurtamohler/31/head -> origin/gh/kurtamohler/31/head 2025-07-17T08:34:51.0248143Z * [new branch] gh/kurtamohler/31/orig -> origin/gh/kurtamohler/31/orig 2025-07-17T08:34:51.0252797Z * [new branch] gh/kurtamohler/32/base -> origin/gh/kurtamohler/32/base 2025-07-17T08:34:51.0256208Z * [new branch] gh/kurtamohler/32/head -> origin/gh/kurtamohler/32/head 2025-07-17T08:34:51.0259702Z * [new branch] gh/kurtamohler/32/orig -> origin/gh/kurtamohler/32/orig 2025-07-17T08:34:51.0264534Z * [new branch] gh/kurtamohler/33/base -> origin/gh/kurtamohler/33/base 2025-07-17T08:34:51.0267678Z * [new branch] gh/kurtamohler/33/head -> origin/gh/kurtamohler/33/head 2025-07-17T08:34:51.0271087Z * [new branch] gh/kurtamohler/33/orig -> origin/gh/kurtamohler/33/orig 2025-07-17T08:34:51.0275868Z * [new branch] gh/kurtamohler/34/base -> origin/gh/kurtamohler/34/base 2025-07-17T08:34:51.0279501Z * [new branch] gh/kurtamohler/34/head -> origin/gh/kurtamohler/34/head 2025-07-17T08:34:51.0282844Z * [new branch] gh/kurtamohler/34/orig -> origin/gh/kurtamohler/34/orig 2025-07-17T08:34:51.0287400Z * [new branch] gh/kurtamohler/37/base -> origin/gh/kurtamohler/37/base 2025-07-17T08:34:51.0290754Z * [new branch] gh/kurtamohler/37/head -> origin/gh/kurtamohler/37/head 2025-07-17T08:34:51.0294099Z * [new branch] gh/kurtamohler/37/orig -> origin/gh/kurtamohler/37/orig 2025-07-17T08:34:51.0298640Z * [new branch] gh/kurtamohler/38/base -> origin/gh/kurtamohler/38/base 2025-07-17T08:34:51.0302053Z * [new branch] gh/kurtamohler/38/head -> origin/gh/kurtamohler/38/head 2025-07-17T08:34:51.0305500Z * [new branch] gh/kurtamohler/38/orig -> origin/gh/kurtamohler/38/orig 2025-07-17T08:34:51.0310072Z * [new branch] gh/kurtamohler/39/base -> origin/gh/kurtamohler/39/base 2025-07-17T08:34:51.0313429Z * [new branch] gh/kurtamohler/39/head -> origin/gh/kurtamohler/39/head 2025-07-17T08:34:51.0316989Z * [new branch] gh/kurtamohler/39/orig -> origin/gh/kurtamohler/39/orig 2025-07-17T08:34:51.0321659Z * [new branch] gh/kurtamohler/40/base -> origin/gh/kurtamohler/40/base 2025-07-17T08:34:51.0324996Z * [new branch] gh/kurtamohler/40/head -> origin/gh/kurtamohler/40/head 2025-07-17T08:34:51.0328332Z * [new branch] gh/kurtamohler/40/orig -> origin/gh/kurtamohler/40/orig 2025-07-17T08:34:51.0332984Z * [new branch] gh/kurtamohler/41/base -> origin/gh/kurtamohler/41/base 2025-07-17T08:34:51.0336334Z * [new branch] gh/kurtamohler/41/head -> origin/gh/kurtamohler/41/head 2025-07-17T08:34:51.0339729Z * [new branch] gh/kurtamohler/41/orig -> origin/gh/kurtamohler/41/orig 2025-07-17T08:34:51.0345341Z * [new branch] gh/kwen2501/1/base -> origin/gh/kwen2501/1/base 2025-07-17T08:34:51.0348686Z * [new branch] gh/kwen2501/1/head -> origin/gh/kwen2501/1/head 2025-07-17T08:34:51.0353549Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-07-17T08:34:51.0357776Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-07-17T08:34:51.0360834Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-07-17T08:34:51.0365314Z * [new branch] gh/kwen2501/131/base -> origin/gh/kwen2501/131/base 2025-07-17T08:34:51.0368689Z * [new branch] gh/kwen2501/131/head -> origin/gh/kwen2501/131/head 2025-07-17T08:34:51.0372058Z * [new branch] gh/kwen2501/131/orig -> origin/gh/kwen2501/131/orig 2025-07-17T08:34:51.0376730Z * [new branch] gh/kwen2501/142/base -> origin/gh/kwen2501/142/base 2025-07-17T08:34:51.0380087Z * [new branch] gh/kwen2501/142/head -> origin/gh/kwen2501/142/head 2025-07-17T08:34:51.0383441Z * [new branch] gh/kwen2501/142/orig -> origin/gh/kwen2501/142/orig 2025-07-17T08:34:51.0388123Z * [new branch] gh/kwen2501/15/base -> origin/gh/kwen2501/15/base 2025-07-17T08:34:51.0391619Z * [new branch] gh/kwen2501/15/head -> origin/gh/kwen2501/15/head 2025-07-17T08:34:51.0396830Z * [new branch] gh/kwen2501/156/base -> origin/gh/kwen2501/156/base 2025-07-17T08:34:51.0400430Z * [new branch] gh/kwen2501/156/head -> origin/gh/kwen2501/156/head 2025-07-17T08:34:51.0403501Z * [new branch] gh/kwen2501/156/orig -> origin/gh/kwen2501/156/orig 2025-07-17T08:34:51.0408059Z * [new branch] gh/kwen2501/162/base -> origin/gh/kwen2501/162/base 2025-07-17T08:34:51.0411470Z * [new branch] gh/kwen2501/162/head -> origin/gh/kwen2501/162/head 2025-07-17T08:34:51.0414832Z * [new branch] gh/kwen2501/162/orig -> origin/gh/kwen2501/162/orig 2025-07-17T08:34:51.0419434Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-07-17T08:34:51.0422744Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-07-17T08:34:51.0427253Z * [new branch] gh/kwen2501/172/base -> origin/gh/kwen2501/172/base 2025-07-17T08:34:51.0430625Z * [new branch] gh/kwen2501/172/head -> origin/gh/kwen2501/172/head 2025-07-17T08:34:51.0434287Z * [new branch] gh/kwen2501/172/orig -> origin/gh/kwen2501/172/orig 2025-07-17T08:34:51.0440472Z * [new branch] gh/kwen2501/173/base -> origin/gh/kwen2501/173/base 2025-07-17T08:34:51.0443698Z * [new branch] gh/kwen2501/173/head -> origin/gh/kwen2501/173/head 2025-07-17T08:34:51.0447075Z * [new branch] gh/kwen2501/173/orig -> origin/gh/kwen2501/173/orig 2025-07-17T08:34:51.0451546Z * [new branch] gh/kwen2501/174/base -> origin/gh/kwen2501/174/base 2025-07-17T08:34:51.0454906Z * [new branch] gh/kwen2501/174/head -> origin/gh/kwen2501/174/head 2025-07-17T08:34:51.0458286Z * [new branch] gh/kwen2501/174/orig -> origin/gh/kwen2501/174/orig 2025-07-17T08:34:51.0462813Z * [new branch] gh/kwen2501/175/base -> origin/gh/kwen2501/175/base 2025-07-17T08:34:51.0466074Z * [new branch] gh/kwen2501/175/head -> origin/gh/kwen2501/175/head 2025-07-17T08:34:51.0469375Z * [new branch] gh/kwen2501/175/orig -> origin/gh/kwen2501/175/orig 2025-07-17T08:34:51.0473862Z * [new branch] gh/kwen2501/176/base -> origin/gh/kwen2501/176/base 2025-07-17T08:34:51.0477412Z * [new branch] gh/kwen2501/176/head -> origin/gh/kwen2501/176/head 2025-07-17T08:34:51.0480717Z * [new branch] gh/kwen2501/176/orig -> origin/gh/kwen2501/176/orig 2025-07-17T08:34:51.0485365Z * [new branch] gh/kwen2501/177/base -> origin/gh/kwen2501/177/base 2025-07-17T08:34:51.0488705Z * [new branch] gh/kwen2501/177/head -> origin/gh/kwen2501/177/head 2025-07-17T08:34:51.0492066Z * [new branch] gh/kwen2501/177/orig -> origin/gh/kwen2501/177/orig 2025-07-17T08:34:51.0502141Z * [new branch] gh/kwen2501/178/base -> origin/gh/kwen2501/178/base 2025-07-17T08:34:51.0502521Z * [new branch] gh/kwen2501/178/head -> origin/gh/kwen2501/178/head 2025-07-17T08:34:51.0503291Z * [new branch] gh/kwen2501/178/orig -> origin/gh/kwen2501/178/orig 2025-07-17T08:34:51.0507959Z * [new branch] gh/kwen2501/179/base -> origin/gh/kwen2501/179/base 2025-07-17T08:34:51.0511233Z * [new branch] gh/kwen2501/179/head -> origin/gh/kwen2501/179/head 2025-07-17T08:34:51.0514736Z * [new branch] gh/kwen2501/179/orig -> origin/gh/kwen2501/179/orig 2025-07-17T08:34:51.0519431Z * [new branch] gh/kwen2501/180/base -> origin/gh/kwen2501/180/base 2025-07-17T08:34:51.0522716Z * [new branch] gh/kwen2501/180/head -> origin/gh/kwen2501/180/head 2025-07-17T08:34:51.0526045Z * [new branch] gh/kwen2501/180/orig -> origin/gh/kwen2501/180/orig 2025-07-17T08:34:51.0530636Z * [new branch] gh/kwen2501/181/base -> origin/gh/kwen2501/181/base 2025-07-17T08:34:51.0534168Z * [new branch] gh/kwen2501/181/head -> origin/gh/kwen2501/181/head 2025-07-17T08:34:51.0537283Z * [new branch] gh/kwen2501/181/orig -> origin/gh/kwen2501/181/orig 2025-07-17T08:34:51.0541789Z * [new branch] gh/kwen2501/182/base -> origin/gh/kwen2501/182/base 2025-07-17T08:34:51.0545098Z * [new branch] gh/kwen2501/182/head -> origin/gh/kwen2501/182/head 2025-07-17T08:34:51.0548391Z * [new branch] gh/kwen2501/182/orig -> origin/gh/kwen2501/182/orig 2025-07-17T08:34:51.0552999Z * [new branch] gh/kwen2501/183/base -> origin/gh/kwen2501/183/base 2025-07-17T08:34:51.0556497Z * [new branch] gh/kwen2501/183/head -> origin/gh/kwen2501/183/head 2025-07-17T08:34:51.0559833Z * [new branch] gh/kwen2501/183/orig -> origin/gh/kwen2501/183/orig 2025-07-17T08:34:51.0564289Z * [new branch] gh/kwen2501/184/base -> origin/gh/kwen2501/184/base 2025-07-17T08:34:51.0567746Z * [new branch] gh/kwen2501/184/head -> origin/gh/kwen2501/184/head 2025-07-17T08:34:51.0571064Z * [new branch] gh/kwen2501/184/orig -> origin/gh/kwen2501/184/orig 2025-07-17T08:34:51.0575710Z * [new branch] gh/kwen2501/185/base -> origin/gh/kwen2501/185/base 2025-07-17T08:34:51.0578978Z * [new branch] gh/kwen2501/185/head -> origin/gh/kwen2501/185/head 2025-07-17T08:34:51.0582297Z * [new branch] gh/kwen2501/185/orig -> origin/gh/kwen2501/185/orig 2025-07-17T08:34:51.0586636Z * [new branch] gh/kwen2501/186/base -> origin/gh/kwen2501/186/base 2025-07-17T08:34:51.0589991Z * [new branch] gh/kwen2501/186/head -> origin/gh/kwen2501/186/head 2025-07-17T08:34:51.0593373Z * [new branch] gh/kwen2501/186/orig -> origin/gh/kwen2501/186/orig 2025-07-17T08:34:51.0598581Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-07-17T08:34:51.0601977Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-07-17T08:34:51.0605366Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-07-17T08:34:51.0609891Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-07-17T08:34:51.0613203Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-07-17T08:34:51.0616479Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-07-17T08:34:51.0621129Z * [new branch] gh/kwen2501/189/base -> origin/gh/kwen2501/189/base 2025-07-17T08:34:51.0624759Z * [new branch] gh/kwen2501/189/head -> origin/gh/kwen2501/189/head 2025-07-17T08:34:51.0628119Z * [new branch] gh/kwen2501/189/orig -> origin/gh/kwen2501/189/orig 2025-07-17T08:34:51.0632642Z * [new branch] gh/kwen2501/190/base -> origin/gh/kwen2501/190/base 2025-07-17T08:34:51.0636109Z * [new branch] gh/kwen2501/190/head -> origin/gh/kwen2501/190/head 2025-07-17T08:34:51.0639443Z * [new branch] gh/kwen2501/190/orig -> origin/gh/kwen2501/190/orig 2025-07-17T08:34:51.0644020Z * [new branch] gh/kwen2501/191/base -> origin/gh/kwen2501/191/base 2025-07-17T08:34:51.0647325Z * [new branch] gh/kwen2501/191/head -> origin/gh/kwen2501/191/head 2025-07-17T08:34:51.0650729Z * [new branch] gh/kwen2501/191/orig -> origin/gh/kwen2501/191/orig 2025-07-17T08:34:51.0655298Z * [new branch] gh/kwen2501/192/base -> origin/gh/kwen2501/192/base 2025-07-17T08:34:51.0658536Z * [new branch] gh/kwen2501/192/head -> origin/gh/kwen2501/192/head 2025-07-17T08:34:51.0661878Z * [new branch] gh/kwen2501/192/orig -> origin/gh/kwen2501/192/orig 2025-07-17T08:34:51.0666620Z * [new branch] gh/kwen2501/193/base -> origin/gh/kwen2501/193/base 2025-07-17T08:34:51.0669816Z * [new branch] gh/kwen2501/193/head -> origin/gh/kwen2501/193/head 2025-07-17T08:34:51.0673136Z * [new branch] gh/kwen2501/193/orig -> origin/gh/kwen2501/193/orig 2025-07-17T08:34:51.0679287Z * [new branch] gh/kwen2501/194/base -> origin/gh/kwen2501/194/base 2025-07-17T08:34:51.0682367Z * [new branch] gh/kwen2501/194/head -> origin/gh/kwen2501/194/head 2025-07-17T08:34:51.0685703Z * [new branch] gh/kwen2501/194/orig -> origin/gh/kwen2501/194/orig 2025-07-17T08:34:51.0690542Z * [new branch] gh/kwen2501/195/base -> origin/gh/kwen2501/195/base 2025-07-17T08:34:51.0693883Z * [new branch] gh/kwen2501/195/head -> origin/gh/kwen2501/195/head 2025-07-17T08:34:51.0697207Z * [new branch] gh/kwen2501/195/orig -> origin/gh/kwen2501/195/orig 2025-07-17T08:34:51.0701688Z * [new branch] gh/kwen2501/196/base -> origin/gh/kwen2501/196/base 2025-07-17T08:34:51.0704997Z * [new branch] gh/kwen2501/196/head -> origin/gh/kwen2501/196/head 2025-07-17T08:34:51.0708298Z * [new branch] gh/kwen2501/196/orig -> origin/gh/kwen2501/196/orig 2025-07-17T08:34:51.0714218Z * [new branch] gh/laithsakka/112/head -> origin/gh/laithsakka/112/head 2025-07-17T08:34:51.0718952Z * [new branch] gh/laithsakka/117/base -> origin/gh/laithsakka/117/base 2025-07-17T08:34:51.0722298Z * [new branch] gh/laithsakka/117/head -> origin/gh/laithsakka/117/head 2025-07-17T08:34:51.0725676Z * [new branch] gh/laithsakka/117/orig -> origin/gh/laithsakka/117/orig 2025-07-17T08:34:51.0730177Z * [new branch] gh/laithsakka/120/base -> origin/gh/laithsakka/120/base 2025-07-17T08:34:51.0733520Z * [new branch] gh/laithsakka/120/head -> origin/gh/laithsakka/120/head 2025-07-17T08:34:51.0736868Z * [new branch] gh/laithsakka/120/orig -> origin/gh/laithsakka/120/orig 2025-07-17T08:34:51.0741518Z * [new branch] gh/laithsakka/141/base -> origin/gh/laithsakka/141/base 2025-07-17T08:34:51.0744839Z * [new branch] gh/laithsakka/141/head -> origin/gh/laithsakka/141/head 2025-07-17T08:34:51.0748254Z * [new branch] gh/laithsakka/141/orig -> origin/gh/laithsakka/141/orig 2025-07-17T08:34:51.0752905Z * [new branch] gh/laithsakka/152/base -> origin/gh/laithsakka/152/base 2025-07-17T08:34:51.0756425Z * [new branch] gh/laithsakka/152/head -> origin/gh/laithsakka/152/head 2025-07-17T08:34:51.0759750Z * [new branch] gh/laithsakka/152/orig -> origin/gh/laithsakka/152/orig 2025-07-17T08:34:51.0764349Z * [new branch] gh/laithsakka/156/base -> origin/gh/laithsakka/156/base 2025-07-17T08:34:51.0767737Z * [new branch] gh/laithsakka/156/head -> origin/gh/laithsakka/156/head 2025-07-17T08:34:51.0771029Z * [new branch] gh/laithsakka/156/orig -> origin/gh/laithsakka/156/orig 2025-07-17T08:34:51.0775490Z * [new branch] gh/laithsakka/159/base -> origin/gh/laithsakka/159/base 2025-07-17T08:34:51.0778797Z * [new branch] gh/laithsakka/159/head -> origin/gh/laithsakka/159/head 2025-07-17T08:34:51.0782117Z * [new branch] gh/laithsakka/159/orig -> origin/gh/laithsakka/159/orig 2025-07-17T08:34:51.0786897Z * [new branch] gh/laithsakka/160/base -> origin/gh/laithsakka/160/base 2025-07-17T08:34:51.0790186Z * [new branch] gh/laithsakka/160/head -> origin/gh/laithsakka/160/head 2025-07-17T08:34:51.0793486Z * [new branch] gh/laithsakka/160/orig -> origin/gh/laithsakka/160/orig 2025-07-17T08:34:51.0798329Z * [new branch] gh/laithsakka/178/base -> origin/gh/laithsakka/178/base 2025-07-17T08:34:51.0801811Z * [new branch] gh/laithsakka/178/head -> origin/gh/laithsakka/178/head 2025-07-17T08:34:51.0805020Z * [new branch] gh/laithsakka/178/orig -> origin/gh/laithsakka/178/orig 2025-07-17T08:34:51.0809533Z * [new branch] gh/laithsakka/191/base -> origin/gh/laithsakka/191/base 2025-07-17T08:34:51.0812848Z * [new branch] gh/laithsakka/191/head -> origin/gh/laithsakka/191/head 2025-07-17T08:34:51.0816453Z * [new branch] gh/laithsakka/191/orig -> origin/gh/laithsakka/191/orig 2025-07-17T08:34:51.0821072Z * [new branch] gh/laithsakka/199/base -> origin/gh/laithsakka/199/base 2025-07-17T08:34:51.0824315Z * [new branch] gh/laithsakka/199/head -> origin/gh/laithsakka/199/head 2025-07-17T08:34:51.0827676Z * [new branch] gh/laithsakka/199/orig -> origin/gh/laithsakka/199/orig 2025-07-17T08:34:51.0832407Z * [new branch] gh/laithsakka/206/base -> origin/gh/laithsakka/206/base 2025-07-17T08:34:51.0836036Z * [new branch] gh/laithsakka/206/head -> origin/gh/laithsakka/206/head 2025-07-17T08:34:51.0839340Z * [new branch] gh/laithsakka/206/orig -> origin/gh/laithsakka/206/orig 2025-07-17T08:34:51.0844126Z * [new branch] gh/laithsakka/210/base -> origin/gh/laithsakka/210/base 2025-07-17T08:34:51.0847528Z * [new branch] gh/laithsakka/210/head -> origin/gh/laithsakka/210/head 2025-07-17T08:34:51.0851000Z * [new branch] gh/laithsakka/210/orig -> origin/gh/laithsakka/210/orig 2025-07-17T08:34:51.0855609Z * [new branch] gh/laithsakka/215/base -> origin/gh/laithsakka/215/base 2025-07-17T08:34:51.0858845Z * [new branch] gh/laithsakka/215/head -> origin/gh/laithsakka/215/head 2025-07-17T08:34:51.0862137Z * [new branch] gh/laithsakka/215/orig -> origin/gh/laithsakka/215/orig 2025-07-17T08:34:51.0866870Z * [new branch] gh/laithsakka/219/base -> origin/gh/laithsakka/219/base 2025-07-17T08:34:51.0870161Z * [new branch] gh/laithsakka/219/head -> origin/gh/laithsakka/219/head 2025-07-17T08:34:51.0873559Z * [new branch] gh/laithsakka/219/orig -> origin/gh/laithsakka/219/orig 2025-07-17T08:34:51.0878425Z * [new branch] gh/laithsakka/220/base -> origin/gh/laithsakka/220/base 2025-07-17T08:34:51.0881748Z * [new branch] gh/laithsakka/220/head -> origin/gh/laithsakka/220/head 2025-07-17T08:34:51.0885091Z * [new branch] gh/laithsakka/220/orig -> origin/gh/laithsakka/220/orig 2025-07-17T08:34:51.0889676Z * [new branch] gh/laithsakka/221/base -> origin/gh/laithsakka/221/base 2025-07-17T08:34:51.0893003Z * [new branch] gh/laithsakka/221/head -> origin/gh/laithsakka/221/head 2025-07-17T08:34:51.0896321Z * [new branch] gh/laithsakka/221/orig -> origin/gh/laithsakka/221/orig 2025-07-17T08:34:51.0900869Z * [new branch] gh/laithsakka/222/base -> origin/gh/laithsakka/222/base 2025-07-17T08:34:51.0904277Z * [new branch] gh/laithsakka/222/head -> origin/gh/laithsakka/222/head 2025-07-17T08:34:51.0907653Z * [new branch] gh/laithsakka/222/orig -> origin/gh/laithsakka/222/orig 2025-07-17T08:34:51.0912090Z * [new branch] gh/laithsakka/223/base -> origin/gh/laithsakka/223/base 2025-07-17T08:34:51.0916093Z * [new branch] gh/laithsakka/223/head -> origin/gh/laithsakka/223/head 2025-07-17T08:34:51.0919445Z * [new branch] gh/laithsakka/223/orig -> origin/gh/laithsakka/223/orig 2025-07-17T08:34:51.0924001Z * [new branch] gh/laithsakka/224/base -> origin/gh/laithsakka/224/base 2025-07-17T08:34:51.0927382Z * [new branch] gh/laithsakka/224/head -> origin/gh/laithsakka/224/head 2025-07-17T08:34:51.0930903Z * [new branch] gh/laithsakka/224/orig -> origin/gh/laithsakka/224/orig 2025-07-17T08:34:51.0935184Z * [new branch] gh/laithsakka/225/base -> origin/gh/laithsakka/225/base 2025-07-17T08:34:51.0938459Z * [new branch] gh/laithsakka/225/head -> origin/gh/laithsakka/225/head 2025-07-17T08:34:51.0941803Z * [new branch] gh/laithsakka/225/orig -> origin/gh/laithsakka/225/orig 2025-07-17T08:34:51.0946169Z * [new branch] gh/laithsakka/226/base -> origin/gh/laithsakka/226/base 2025-07-17T08:34:51.0949420Z * [new branch] gh/laithsakka/226/head -> origin/gh/laithsakka/226/head 2025-07-17T08:34:51.0952783Z * [new branch] gh/laithsakka/226/orig -> origin/gh/laithsakka/226/orig 2025-07-17T08:34:51.0958887Z * [new branch] gh/laithsakka/227/base -> origin/gh/laithsakka/227/base 2025-07-17T08:34:51.0962144Z * [new branch] gh/laithsakka/227/head -> origin/gh/laithsakka/227/head 2025-07-17T08:34:51.0965442Z * [new branch] gh/laithsakka/227/orig -> origin/gh/laithsakka/227/orig 2025-07-17T08:34:51.0970203Z * [new branch] gh/laithsakka/228/base -> origin/gh/laithsakka/228/base 2025-07-17T08:34:51.0973501Z * [new branch] gh/laithsakka/228/head -> origin/gh/laithsakka/228/head 2025-07-17T08:34:51.0976800Z * [new branch] gh/laithsakka/228/orig -> origin/gh/laithsakka/228/orig 2025-07-17T08:34:51.0981280Z * [new branch] gh/laithsakka/229/base -> origin/gh/laithsakka/229/base 2025-07-17T08:34:51.0984622Z * [new branch] gh/laithsakka/229/head -> origin/gh/laithsakka/229/head 2025-07-17T08:34:51.0987902Z * [new branch] gh/laithsakka/229/orig -> origin/gh/laithsakka/229/orig 2025-07-17T08:34:51.0992327Z * [new branch] gh/laithsakka/230/base -> origin/gh/laithsakka/230/base 2025-07-17T08:34:51.0995871Z * [new branch] gh/laithsakka/230/head -> origin/gh/laithsakka/230/head 2025-07-17T08:34:51.0999202Z * [new branch] gh/laithsakka/230/orig -> origin/gh/laithsakka/230/orig 2025-07-17T08:34:51.1003664Z * [new branch] gh/laithsakka/231/base -> origin/gh/laithsakka/231/base 2025-07-17T08:34:51.1007211Z * [new branch] gh/laithsakka/231/head -> origin/gh/laithsakka/231/head 2025-07-17T08:34:51.1010542Z * [new branch] gh/laithsakka/231/orig -> origin/gh/laithsakka/231/orig 2025-07-17T08:34:51.1015273Z * [new branch] gh/laithsakka/232/base -> origin/gh/laithsakka/232/base 2025-07-17T08:34:51.1018629Z * [new branch] gh/laithsakka/232/head -> origin/gh/laithsakka/232/head 2025-07-17T08:34:51.1021987Z * [new branch] gh/laithsakka/232/orig -> origin/gh/laithsakka/232/orig 2025-07-17T08:34:51.1026771Z * [new branch] gh/laithsakka/233/base -> origin/gh/laithsakka/233/base 2025-07-17T08:34:51.1030098Z * [new branch] gh/laithsakka/233/head -> origin/gh/laithsakka/233/head 2025-07-17T08:34:51.1033474Z * [new branch] gh/laithsakka/233/orig -> origin/gh/laithsakka/233/orig 2025-07-17T08:34:51.1038188Z * [new branch] gh/laithsakka/234/base -> origin/gh/laithsakka/234/base 2025-07-17T08:34:51.1041442Z * [new branch] gh/laithsakka/234/head -> origin/gh/laithsakka/234/head 2025-07-17T08:34:51.1044784Z * [new branch] gh/laithsakka/234/orig -> origin/gh/laithsakka/234/orig 2025-07-17T08:34:51.1049266Z * [new branch] gh/laithsakka/235/base -> origin/gh/laithsakka/235/base 2025-07-17T08:34:51.1052668Z * [new branch] gh/laithsakka/235/head -> origin/gh/laithsakka/235/head 2025-07-17T08:34:51.1055985Z * [new branch] gh/laithsakka/235/orig -> origin/gh/laithsakka/235/orig 2025-07-17T08:34:51.1060711Z * [new branch] gh/laithsakka/236/base -> origin/gh/laithsakka/236/base 2025-07-17T08:34:51.1063853Z * [new branch] gh/laithsakka/236/head -> origin/gh/laithsakka/236/head 2025-07-17T08:34:51.1067170Z * [new branch] gh/laithsakka/236/orig -> origin/gh/laithsakka/236/orig 2025-07-17T08:34:51.1071831Z * [new branch] gh/laithsakka/237/base -> origin/gh/laithsakka/237/base 2025-07-17T08:34:51.1075314Z * [new branch] gh/laithsakka/237/head -> origin/gh/laithsakka/237/head 2025-07-17T08:34:51.1078632Z * [new branch] gh/laithsakka/237/orig -> origin/gh/laithsakka/237/orig 2025-07-17T08:34:51.1083068Z * [new branch] gh/laithsakka/238/base -> origin/gh/laithsakka/238/base 2025-07-17T08:34:51.1086376Z * [new branch] gh/laithsakka/238/head -> origin/gh/laithsakka/238/head 2025-07-17T08:34:51.1089684Z * [new branch] gh/laithsakka/238/orig -> origin/gh/laithsakka/238/orig 2025-07-17T08:34:51.1094194Z * [new branch] gh/laithsakka/239/base -> origin/gh/laithsakka/239/base 2025-07-17T08:34:51.1097490Z * [new branch] gh/laithsakka/239/head -> origin/gh/laithsakka/239/head 2025-07-17T08:34:51.1100802Z * [new branch] gh/laithsakka/239/orig -> origin/gh/laithsakka/239/orig 2025-07-17T08:34:51.1105349Z * [new branch] gh/laithsakka/240/base -> origin/gh/laithsakka/240/base 2025-07-17T08:34:51.1108666Z * [new branch] gh/laithsakka/240/head -> origin/gh/laithsakka/240/head 2025-07-17T08:34:51.1112129Z * [new branch] gh/laithsakka/240/orig -> origin/gh/laithsakka/240/orig 2025-07-17T08:34:51.1116976Z * [new branch] gh/laithsakka/241/base -> origin/gh/laithsakka/241/base 2025-07-17T08:34:51.1120221Z * [new branch] gh/laithsakka/241/head -> origin/gh/laithsakka/241/head 2025-07-17T08:34:51.1123542Z * [new branch] gh/laithsakka/241/orig -> origin/gh/laithsakka/241/orig 2025-07-17T08:34:51.1128225Z * [new branch] gh/laithsakka/242/base -> origin/gh/laithsakka/242/base 2025-07-17T08:34:51.1131572Z * [new branch] gh/laithsakka/242/head -> origin/gh/laithsakka/242/head 2025-07-17T08:34:51.1134878Z * [new branch] gh/laithsakka/242/orig -> origin/gh/laithsakka/242/orig 2025-07-17T08:34:51.1139314Z * [new branch] gh/laithsakka/243/base -> origin/gh/laithsakka/243/base 2025-07-17T08:34:51.1142612Z * [new branch] gh/laithsakka/243/head -> origin/gh/laithsakka/243/head 2025-07-17T08:34:51.1145932Z * [new branch] gh/laithsakka/243/orig -> origin/gh/laithsakka/243/orig 2025-07-17T08:34:51.1150657Z * [new branch] gh/laithsakka/244/base -> origin/gh/laithsakka/244/base 2025-07-17T08:34:51.1153985Z * [new branch] gh/laithsakka/244/head -> origin/gh/laithsakka/244/head 2025-07-17T08:34:51.1157458Z * [new branch] gh/laithsakka/244/orig -> origin/gh/laithsakka/244/orig 2025-07-17T08:34:51.1162202Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-07-17T08:34:51.1166577Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-07-17T08:34:51.1171038Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-07-17T08:34:51.1174417Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-07-17T08:34:51.1178795Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-07-17T08:34:51.1182065Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-07-17T08:34:51.1186433Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-07-17T08:34:51.1189688Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-07-17T08:34:51.1195665Z * [new branch] gh/leslie-fang-intel/190/base -> origin/gh/leslie-fang-intel/190/base 2025-07-17T08:34:51.1199145Z * [new branch] gh/leslie-fang-intel/190/head -> origin/gh/leslie-fang-intel/190/head 2025-07-17T08:34:51.1202587Z * [new branch] gh/leslie-fang-intel/190/orig -> origin/gh/leslie-fang-intel/190/orig 2025-07-17T08:34:51.1207147Z * [new branch] gh/leslie-fang-intel/199/base -> origin/gh/leslie-fang-intel/199/base 2025-07-17T08:34:51.1210456Z * [new branch] gh/leslie-fang-intel/199/head -> origin/gh/leslie-fang-intel/199/head 2025-07-17T08:34:51.1213760Z * [new branch] gh/leslie-fang-intel/199/orig -> origin/gh/leslie-fang-intel/199/orig 2025-07-17T08:34:51.1218289Z * [new branch] gh/leslie-fang-intel/200/base -> origin/gh/leslie-fang-intel/200/base 2025-07-17T08:34:51.1221558Z * [new branch] gh/leslie-fang-intel/200/head -> origin/gh/leslie-fang-intel/200/head 2025-07-17T08:34:51.1224886Z * [new branch] gh/leslie-fang-intel/200/orig -> origin/gh/leslie-fang-intel/200/orig 2025-07-17T08:34:51.1229321Z * [new branch] gh/leslie-fang-intel/201/base -> origin/gh/leslie-fang-intel/201/base 2025-07-17T08:34:51.1232698Z * [new branch] gh/leslie-fang-intel/201/head -> origin/gh/leslie-fang-intel/201/head 2025-07-17T08:34:51.1236224Z * [new branch] gh/leslie-fang-intel/201/orig -> origin/gh/leslie-fang-intel/201/orig 2025-07-17T08:34:51.1240755Z * [new branch] gh/leslie-fang-intel/202/base -> origin/gh/leslie-fang-intel/202/base 2025-07-17T08:34:51.1244070Z * [new branch] gh/leslie-fang-intel/202/head -> origin/gh/leslie-fang-intel/202/head 2025-07-17T08:34:51.1247522Z * [new branch] gh/leslie-fang-intel/202/orig -> origin/gh/leslie-fang-intel/202/orig 2025-07-17T08:34:51.1251999Z * [new branch] gh/leslie-fang-intel/203/base -> origin/gh/leslie-fang-intel/203/base 2025-07-17T08:34:51.1255337Z * [new branch] gh/leslie-fang-intel/203/head -> origin/gh/leslie-fang-intel/203/head 2025-07-17T08:34:51.1258764Z * [new branch] gh/leslie-fang-intel/203/orig -> origin/gh/leslie-fang-intel/203/orig 2025-07-17T08:34:51.1263327Z * [new branch] gh/leslie-fang-intel/204/base -> origin/gh/leslie-fang-intel/204/base 2025-07-17T08:34:51.1266634Z * [new branch] gh/leslie-fang-intel/204/head -> origin/gh/leslie-fang-intel/204/head 2025-07-17T08:34:51.1270006Z * [new branch] gh/leslie-fang-intel/204/orig -> origin/gh/leslie-fang-intel/204/orig 2025-07-17T08:34:51.1274730Z * [new branch] gh/leslie-fang-intel/205/base -> origin/gh/leslie-fang-intel/205/base 2025-07-17T08:34:51.1278119Z * [new branch] gh/leslie-fang-intel/205/head -> origin/gh/leslie-fang-intel/205/head 2025-07-17T08:34:51.1281408Z * [new branch] gh/leslie-fang-intel/205/orig -> origin/gh/leslie-fang-intel/205/orig 2025-07-17T08:34:51.1285910Z * [new branch] gh/leslie-fang-intel/206/base -> origin/gh/leslie-fang-intel/206/base 2025-07-17T08:34:51.1289202Z * [new branch] gh/leslie-fang-intel/206/head -> origin/gh/leslie-fang-intel/206/head 2025-07-17T08:34:51.1292643Z * [new branch] gh/leslie-fang-intel/206/orig -> origin/gh/leslie-fang-intel/206/orig 2025-07-17T08:34:51.1297198Z * [new branch] gh/leslie-fang-intel/207/base -> origin/gh/leslie-fang-intel/207/base 2025-07-17T08:34:51.1300497Z * [new branch] gh/leslie-fang-intel/207/head -> origin/gh/leslie-fang-intel/207/head 2025-07-17T08:34:51.1303849Z * [new branch] gh/leslie-fang-intel/207/orig -> origin/gh/leslie-fang-intel/207/orig 2025-07-17T08:34:51.1308334Z * [new branch] gh/leslie-fang-intel/208/base -> origin/gh/leslie-fang-intel/208/base 2025-07-17T08:34:51.1311654Z * [new branch] gh/leslie-fang-intel/208/head -> origin/gh/leslie-fang-intel/208/head 2025-07-17T08:34:51.1315326Z * [new branch] gh/leslie-fang-intel/208/orig -> origin/gh/leslie-fang-intel/208/orig 2025-07-17T08:34:51.1320626Z * [new branch] gh/lw/1/base -> origin/gh/lw/1/base 2025-07-17T08:34:51.1323867Z * [new branch] gh/lw/1/head -> origin/gh/lw/1/head 2025-07-17T08:34:51.1327197Z * [new branch] gh/lw/1/orig -> origin/gh/lw/1/orig 2025-07-17T08:34:51.1331694Z * [new branch] gh/lw/2/base -> origin/gh/lw/2/base 2025-07-17T08:34:51.1335179Z * [new branch] gh/lw/2/head -> origin/gh/lw/2/head 2025-07-17T08:34:51.1338427Z * [new branch] gh/lw/2/orig -> origin/gh/lw/2/orig 2025-07-17T08:34:51.1344025Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-07-17T08:34:51.1348511Z * [new branch] gh/malfet/169/base -> origin/gh/malfet/169/base 2025-07-17T08:34:51.1351809Z * [new branch] gh/malfet/169/head -> origin/gh/malfet/169/head 2025-07-17T08:34:51.1355977Z * [new branch] gh/malfet/169/orig -> origin/gh/malfet/169/orig 2025-07-17T08:34:51.1360545Z * [new branch] gh/malfet/259/base -> origin/gh/malfet/259/base 2025-07-17T08:34:51.1363787Z * [new branch] gh/malfet/259/head -> origin/gh/malfet/259/head 2025-07-17T08:34:51.1367154Z * [new branch] gh/malfet/259/orig -> origin/gh/malfet/259/orig 2025-07-17T08:34:51.1371693Z * [new branch] gh/malfet/330/base -> origin/gh/malfet/330/base 2025-07-17T08:34:51.1375033Z * [new branch] gh/malfet/330/head -> origin/gh/malfet/330/head 2025-07-17T08:34:51.1378569Z * [new branch] gh/malfet/330/orig -> origin/gh/malfet/330/orig 2025-07-17T08:34:51.1383097Z * [new branch] gh/malfet/394/base -> origin/gh/malfet/394/base 2025-07-17T08:34:51.1386425Z * [new branch] gh/malfet/394/head -> origin/gh/malfet/394/head 2025-07-17T08:34:51.1389928Z * [new branch] gh/malfet/394/orig -> origin/gh/malfet/394/orig 2025-07-17T08:34:51.1395785Z * [new branch] gh/malfet/396/base -> origin/gh/malfet/396/base 2025-07-17T08:34:51.1399221Z * [new branch] gh/malfet/396/head -> origin/gh/malfet/396/head 2025-07-17T08:34:51.1402493Z * [new branch] gh/malfet/396/orig -> origin/gh/malfet/396/orig 2025-07-17T08:34:51.1407250Z * [new branch] gh/malfet/397/base -> origin/gh/malfet/397/base 2025-07-17T08:34:51.1410558Z * [new branch] gh/malfet/397/head -> origin/gh/malfet/397/head 2025-07-17T08:34:51.1413921Z * [new branch] gh/malfet/397/orig -> origin/gh/malfet/397/orig 2025-07-17T08:34:51.1418523Z * [new branch] gh/malfet/398/base -> origin/gh/malfet/398/base 2025-07-17T08:34:51.1421820Z * [new branch] gh/malfet/398/head -> origin/gh/malfet/398/head 2025-07-17T08:34:51.1425246Z * [new branch] gh/malfet/398/orig -> origin/gh/malfet/398/orig 2025-07-17T08:34:51.1429724Z * [new branch] gh/malfet/399/base -> origin/gh/malfet/399/base 2025-07-17T08:34:51.1433021Z * [new branch] gh/malfet/399/head -> origin/gh/malfet/399/head 2025-07-17T08:34:51.1436484Z * [new branch] gh/malfet/399/orig -> origin/gh/malfet/399/orig 2025-07-17T08:34:51.1440950Z * [new branch] gh/malfet/403/base -> origin/gh/malfet/403/base 2025-07-17T08:34:51.1444284Z * [new branch] gh/malfet/403/head -> origin/gh/malfet/403/head 2025-07-17T08:34:51.1447687Z * [new branch] gh/malfet/403/orig -> origin/gh/malfet/403/orig 2025-07-17T08:34:51.1452210Z * [new branch] gh/malfet/404/base -> origin/gh/malfet/404/base 2025-07-17T08:34:51.1455671Z * [new branch] gh/malfet/404/head -> origin/gh/malfet/404/head 2025-07-17T08:34:51.1458862Z * [new branch] gh/malfet/404/orig -> origin/gh/malfet/404/orig 2025-07-17T08:34:51.1463400Z * [new branch] gh/malfet/405/base -> origin/gh/malfet/405/base 2025-07-17T08:34:51.1466635Z * [new branch] gh/malfet/405/head -> origin/gh/malfet/405/head 2025-07-17T08:34:51.1470060Z * [new branch] gh/malfet/405/orig -> origin/gh/malfet/405/orig 2025-07-17T08:34:51.1475132Z * [new branch] gh/malfet/407/base -> origin/gh/malfet/407/base 2025-07-17T08:34:51.1478505Z * [new branch] gh/malfet/407/head -> origin/gh/malfet/407/head 2025-07-17T08:34:51.1481888Z * [new branch] gh/malfet/407/orig -> origin/gh/malfet/407/orig 2025-07-17T08:34:51.1486397Z * [new branch] gh/malfet/408/base -> origin/gh/malfet/408/base 2025-07-17T08:34:51.1489845Z * [new branch] gh/malfet/408/head -> origin/gh/malfet/408/head 2025-07-17T08:34:51.1493166Z * [new branch] gh/malfet/408/orig -> origin/gh/malfet/408/orig 2025-07-17T08:34:51.1497737Z * [new branch] gh/malfet/409/base -> origin/gh/malfet/409/base 2025-07-17T08:34:51.1501051Z * [new branch] gh/malfet/409/head -> origin/gh/malfet/409/head 2025-07-17T08:34:51.1504405Z * [new branch] gh/malfet/409/orig -> origin/gh/malfet/409/orig 2025-07-17T08:34:51.1508994Z * [new branch] gh/malfet/410/base -> origin/gh/malfet/410/base 2025-07-17T08:34:51.1512293Z * [new branch] gh/malfet/410/head -> origin/gh/malfet/410/head 2025-07-17T08:34:51.1515857Z * [new branch] gh/malfet/410/orig -> origin/gh/malfet/410/orig 2025-07-17T08:34:51.1520360Z * [new branch] gh/malfet/411/base -> origin/gh/malfet/411/base 2025-07-17T08:34:51.1523633Z * [new branch] gh/malfet/411/head -> origin/gh/malfet/411/head 2025-07-17T08:34:51.1526915Z * [new branch] gh/malfet/411/orig -> origin/gh/malfet/411/orig 2025-07-17T08:34:51.1531454Z * [new branch] gh/malfet/412/base -> origin/gh/malfet/412/base 2025-07-17T08:34:51.1534742Z * [new branch] gh/malfet/412/head -> origin/gh/malfet/412/head 2025-07-17T08:34:51.1538130Z * [new branch] gh/malfet/412/orig -> origin/gh/malfet/412/orig 2025-07-17T08:34:51.1542623Z * [new branch] gh/malfet/413/base -> origin/gh/malfet/413/base 2025-07-17T08:34:51.1545955Z * [new branch] gh/malfet/413/head -> origin/gh/malfet/413/head 2025-07-17T08:34:51.1549298Z * [new branch] gh/malfet/413/orig -> origin/gh/malfet/413/orig 2025-07-17T08:34:51.1553755Z * [new branch] gh/malfet/414/base -> origin/gh/malfet/414/base 2025-07-17T08:34:51.1557274Z * [new branch] gh/malfet/414/head -> origin/gh/malfet/414/head 2025-07-17T08:34:51.1560697Z * [new branch] gh/malfet/414/orig -> origin/gh/malfet/414/orig 2025-07-17T08:34:51.1565227Z * [new branch] gh/malfet/415/base -> origin/gh/malfet/415/base 2025-07-17T08:34:51.1568542Z * [new branch] gh/malfet/415/head -> origin/gh/malfet/415/head 2025-07-17T08:34:51.1571953Z * [new branch] gh/malfet/415/orig -> origin/gh/malfet/415/orig 2025-07-17T08:34:51.1576486Z * [new branch] gh/malfet/416/base -> origin/gh/malfet/416/base 2025-07-17T08:34:51.1579990Z * [new branch] gh/malfet/416/head -> origin/gh/malfet/416/head 2025-07-17T08:34:51.1583443Z * [new branch] gh/malfet/416/orig -> origin/gh/malfet/416/orig 2025-07-17T08:34:51.1588197Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-07-17T08:34:51.1591356Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-07-17T08:34:51.1594800Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-07-17T08:34:51.1600420Z * [new branch] gh/malfet/418/base -> origin/gh/malfet/418/base 2025-07-17T08:34:51.1604180Z * [new branch] gh/malfet/418/head -> origin/gh/malfet/418/head 2025-07-17T08:34:51.1607530Z * [new branch] gh/malfet/418/orig -> origin/gh/malfet/418/orig 2025-07-17T08:34:51.1612095Z * [new branch] gh/malfet/419/base -> origin/gh/malfet/419/base 2025-07-17T08:34:51.1615449Z * [new branch] gh/malfet/419/head -> origin/gh/malfet/419/head 2025-07-17T08:34:51.1618736Z * [new branch] gh/malfet/419/orig -> origin/gh/malfet/419/orig 2025-07-17T08:34:51.1623235Z * [new branch] gh/malfet/420/base -> origin/gh/malfet/420/base 2025-07-17T08:34:51.1626573Z * [new branch] gh/malfet/420/head -> origin/gh/malfet/420/head 2025-07-17T08:34:51.1630099Z * [new branch] gh/malfet/420/orig -> origin/gh/malfet/420/orig 2025-07-17T08:34:51.1634793Z * [new branch] gh/malfet/421/base -> origin/gh/malfet/421/base 2025-07-17T08:34:51.1638305Z * [new branch] gh/malfet/421/head -> origin/gh/malfet/421/head 2025-07-17T08:34:51.1641661Z * [new branch] gh/malfet/421/orig -> origin/gh/malfet/421/orig 2025-07-17T08:34:51.1646193Z * [new branch] gh/malfet/422/base -> origin/gh/malfet/422/base 2025-07-17T08:34:51.1649553Z * [new branch] gh/malfet/422/head -> origin/gh/malfet/422/head 2025-07-17T08:34:51.1652903Z * [new branch] gh/malfet/422/orig -> origin/gh/malfet/422/orig 2025-07-17T08:34:51.1657506Z * [new branch] gh/malfet/423/base -> origin/gh/malfet/423/base 2025-07-17T08:34:51.1660768Z * [new branch] gh/malfet/423/head -> origin/gh/malfet/423/head 2025-07-17T08:34:51.1664079Z * [new branch] gh/malfet/423/orig -> origin/gh/malfet/423/orig 2025-07-17T08:34:51.1668669Z * [new branch] gh/malfet/424/base -> origin/gh/malfet/424/base 2025-07-17T08:34:51.1672049Z * [new branch] gh/malfet/424/head -> origin/gh/malfet/424/head 2025-07-17T08:34:51.1675698Z * [new branch] gh/malfet/424/orig -> origin/gh/malfet/424/orig 2025-07-17T08:34:51.1680167Z * [new branch] gh/malfet/425/base -> origin/gh/malfet/425/base 2025-07-17T08:34:51.1683472Z * [new branch] gh/malfet/425/head -> origin/gh/malfet/425/head 2025-07-17T08:34:51.1686811Z * [new branch] gh/malfet/425/orig -> origin/gh/malfet/425/orig 2025-07-17T08:34:51.1691407Z * [new branch] gh/malfet/426/base -> origin/gh/malfet/426/base 2025-07-17T08:34:51.1694869Z * [new branch] gh/malfet/426/head -> origin/gh/malfet/426/head 2025-07-17T08:34:51.1698300Z * [new branch] gh/malfet/426/orig -> origin/gh/malfet/426/orig 2025-07-17T08:34:51.1702883Z * [new branch] gh/malfet/427/base -> origin/gh/malfet/427/base 2025-07-17T08:34:51.1706237Z * [new branch] gh/malfet/427/head -> origin/gh/malfet/427/head 2025-07-17T08:34:51.1709691Z * [new branch] gh/malfet/427/orig -> origin/gh/malfet/427/orig 2025-07-17T08:34:51.1714354Z * [new branch] gh/malfet/428/base -> origin/gh/malfet/428/base 2025-07-17T08:34:51.1718276Z * [new branch] gh/malfet/428/head -> origin/gh/malfet/428/head 2025-07-17T08:34:51.1721726Z * [new branch] gh/malfet/428/orig -> origin/gh/malfet/428/orig 2025-07-17T08:34:51.1726444Z * [new branch] gh/malfet/429/base -> origin/gh/malfet/429/base 2025-07-17T08:34:51.1729784Z * [new branch] gh/malfet/429/head -> origin/gh/malfet/429/head 2025-07-17T08:34:51.1733109Z * [new branch] gh/malfet/429/orig -> origin/gh/malfet/429/orig 2025-07-17T08:34:51.1737627Z * [new branch] gh/malfet/430/base -> origin/gh/malfet/430/base 2025-07-17T08:34:51.1740975Z * [new branch] gh/malfet/430/head -> origin/gh/malfet/430/head 2025-07-17T08:34:51.1744375Z * [new branch] gh/malfet/430/orig -> origin/gh/malfet/430/orig 2025-07-17T08:34:51.1748923Z * [new branch] gh/malfet/431/base -> origin/gh/malfet/431/base 2025-07-17T08:34:51.1752220Z * [new branch] gh/malfet/431/head -> origin/gh/malfet/431/head 2025-07-17T08:34:51.1755773Z * [new branch] gh/malfet/431/orig -> origin/gh/malfet/431/orig 2025-07-17T08:34:51.1760455Z * [new branch] gh/malfet/432/base -> origin/gh/malfet/432/base 2025-07-17T08:34:51.1763694Z * [new branch] gh/malfet/432/head -> origin/gh/malfet/432/head 2025-07-17T08:34:51.1767376Z * [new branch] gh/malfet/432/orig -> origin/gh/malfet/432/orig 2025-07-17T08:34:51.1771968Z * [new branch] gh/malfet/433/base -> origin/gh/malfet/433/base 2025-07-17T08:34:51.1775298Z * [new branch] gh/malfet/433/head -> origin/gh/malfet/433/head 2025-07-17T08:34:51.1778722Z * [new branch] gh/malfet/433/orig -> origin/gh/malfet/433/orig 2025-07-17T08:34:51.1783353Z * [new branch] gh/malfet/434/base -> origin/gh/malfet/434/base 2025-07-17T08:34:51.1786658Z * [new branch] gh/malfet/434/head -> origin/gh/malfet/434/head 2025-07-17T08:34:51.1790015Z * [new branch] gh/malfet/434/orig -> origin/gh/malfet/434/orig 2025-07-17T08:34:51.1794979Z * [new branch] gh/malfet/435/base -> origin/gh/malfet/435/base 2025-07-17T08:34:51.1798388Z * [new branch] gh/malfet/435/head -> origin/gh/malfet/435/head 2025-07-17T08:34:51.1801713Z * [new branch] gh/malfet/435/orig -> origin/gh/malfet/435/orig 2025-07-17T08:34:51.1806204Z * [new branch] gh/malfet/436/base -> origin/gh/malfet/436/base 2025-07-17T08:34:51.1809639Z * [new branch] gh/malfet/436/head -> origin/gh/malfet/436/head 2025-07-17T08:34:51.1813125Z * [new branch] gh/malfet/436/orig -> origin/gh/malfet/436/orig 2025-07-17T08:34:51.1817515Z * [new branch] gh/malfet/437/base -> origin/gh/malfet/437/base 2025-07-17T08:34:51.1820855Z * [new branch] gh/malfet/437/head -> origin/gh/malfet/437/head 2025-07-17T08:34:51.1824153Z * [new branch] gh/malfet/437/orig -> origin/gh/malfet/437/orig 2025-07-17T08:34:51.1828676Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-07-17T08:34:51.1831989Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-07-17T08:34:51.1837909Z * [new branch] gh/manuelcandales/10/base -> origin/gh/manuelcandales/10/base 2025-07-17T08:34:51.1841192Z * [new branch] gh/manuelcandales/10/head -> origin/gh/manuelcandales/10/head 2025-07-17T08:34:51.1844505Z * [new branch] gh/manuelcandales/10/orig -> origin/gh/manuelcandales/10/orig 2025-07-17T08:34:51.1849099Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-07-17T08:34:51.1852415Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-07-17T08:34:51.1856004Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-07-17T08:34:51.1860793Z * [new branch] gh/manuelcandales/12/base -> origin/gh/manuelcandales/12/base 2025-07-17T08:34:51.1863896Z * [new branch] gh/manuelcandales/12/head -> origin/gh/manuelcandales/12/head 2025-07-17T08:34:51.1867297Z * [new branch] gh/manuelcandales/12/orig -> origin/gh/manuelcandales/12/orig 2025-07-17T08:34:51.1871769Z * [new branch] gh/manuelcandales/13/base -> origin/gh/manuelcandales/13/base 2025-07-17T08:34:51.1875242Z * [new branch] gh/manuelcandales/13/head -> origin/gh/manuelcandales/13/head 2025-07-17T08:34:51.1878697Z * [new branch] gh/manuelcandales/13/orig -> origin/gh/manuelcandales/13/orig 2025-07-17T08:34:51.1883203Z * [new branch] gh/manuelcandales/14/base -> origin/gh/manuelcandales/14/base 2025-07-17T08:34:51.1886511Z * [new branch] gh/manuelcandales/14/head -> origin/gh/manuelcandales/14/head 2025-07-17T08:34:51.1889834Z * [new branch] gh/manuelcandales/14/orig -> origin/gh/manuelcandales/14/orig 2025-07-17T08:34:51.1894414Z * [new branch] gh/manuelcandales/15/base -> origin/gh/manuelcandales/15/base 2025-07-17T08:34:51.1897760Z * [new branch] gh/manuelcandales/15/head -> origin/gh/manuelcandales/15/head 2025-07-17T08:34:51.1901087Z * [new branch] gh/manuelcandales/15/orig -> origin/gh/manuelcandales/15/orig 2025-07-17T08:34:51.1905599Z * [new branch] gh/manuelcandales/16/base -> origin/gh/manuelcandales/16/base 2025-07-17T08:34:51.1908894Z * [new branch] gh/manuelcandales/16/head -> origin/gh/manuelcandales/16/head 2025-07-17T08:34:51.1912164Z * [new branch] gh/manuelcandales/16/orig -> origin/gh/manuelcandales/16/orig 2025-07-17T08:34:51.1917046Z * [new branch] gh/manuelcandales/17/base -> origin/gh/manuelcandales/17/base 2025-07-17T08:34:51.1920334Z * [new branch] gh/manuelcandales/17/head -> origin/gh/manuelcandales/17/head 2025-07-17T08:34:51.1923672Z * [new branch] gh/manuelcandales/17/orig -> origin/gh/manuelcandales/17/orig 2025-07-17T08:34:51.1928159Z * [new branch] gh/manuelcandales/8/base -> origin/gh/manuelcandales/8/base 2025-07-17T08:34:51.1931450Z * [new branch] gh/manuelcandales/8/head -> origin/gh/manuelcandales/8/head 2025-07-17T08:34:51.1934783Z * [new branch] gh/manuelcandales/8/orig -> origin/gh/manuelcandales/8/orig 2025-07-17T08:34:51.1939228Z * [new branch] gh/manuelcandales/9/base -> origin/gh/manuelcandales/9/base 2025-07-17T08:34:51.1942575Z * [new branch] gh/manuelcandales/9/head -> origin/gh/manuelcandales/9/head 2025-07-17T08:34:51.1946106Z * [new branch] gh/manuelcandales/9/orig -> origin/gh/manuelcandales/9/orig 2025-07-17T08:34:51.1951732Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-07-17T08:34:51.1957566Z * [new branch] gh/masnesral/204/base -> origin/gh/masnesral/204/base 2025-07-17T08:34:51.1960895Z * [new branch] gh/masnesral/204/head -> origin/gh/masnesral/204/head 2025-07-17T08:34:51.1964534Z * [new branch] gh/masnesral/204/orig -> origin/gh/masnesral/204/orig 2025-07-17T08:34:51.1969116Z * [new branch] gh/masnesral/208/base -> origin/gh/masnesral/208/base 2025-07-17T08:34:51.1972551Z * [new branch] gh/masnesral/208/head -> origin/gh/masnesral/208/head 2025-07-17T08:34:51.1975878Z * [new branch] gh/masnesral/208/orig -> origin/gh/masnesral/208/orig 2025-07-17T08:34:51.1980672Z * [new branch] gh/masnesral/209/base -> origin/gh/masnesral/209/base 2025-07-17T08:34:51.1984066Z * [new branch] gh/masnesral/209/head -> origin/gh/masnesral/209/head 2025-07-17T08:34:51.1987543Z * [new branch] gh/masnesral/209/orig -> origin/gh/masnesral/209/orig 2025-07-17T08:34:51.1992535Z * [new branch] gh/masnesral/210/base -> origin/gh/masnesral/210/base 2025-07-17T08:34:51.1996149Z * [new branch] gh/masnesral/210/head -> origin/gh/masnesral/210/head 2025-07-17T08:34:51.1999511Z * [new branch] gh/masnesral/210/orig -> origin/gh/masnesral/210/orig 2025-07-17T08:34:51.2004143Z * [new branch] gh/masnesral/211/base -> origin/gh/masnesral/211/base 2025-07-17T08:34:51.2007455Z * [new branch] gh/masnesral/211/head -> origin/gh/masnesral/211/head 2025-07-17T08:34:51.2010820Z * [new branch] gh/masnesral/211/orig -> origin/gh/masnesral/211/orig 2025-07-17T08:34:51.2015405Z * [new branch] gh/masnesral/212/base -> origin/gh/masnesral/212/base 2025-07-17T08:34:51.2018798Z * [new branch] gh/masnesral/212/head -> origin/gh/masnesral/212/head 2025-07-17T08:34:51.2022191Z * [new branch] gh/masnesral/212/orig -> origin/gh/masnesral/212/orig 2025-07-17T08:34:51.2026934Z * [new branch] gh/masnesral/213/base -> origin/gh/masnesral/213/base 2025-07-17T08:34:51.2030226Z * [new branch] gh/masnesral/213/head -> origin/gh/masnesral/213/head 2025-07-17T08:34:51.2033686Z * [new branch] gh/masnesral/213/orig -> origin/gh/masnesral/213/orig 2025-07-17T08:34:51.2038618Z * [new branch] gh/masnesral/214/base -> origin/gh/masnesral/214/base 2025-07-17T08:34:51.2041930Z * [new branch] gh/masnesral/214/head -> origin/gh/masnesral/214/head 2025-07-17T08:34:51.2045262Z * [new branch] gh/masnesral/214/orig -> origin/gh/masnesral/214/orig 2025-07-17T08:34:51.2049826Z * [new branch] gh/masnesral/215/base -> origin/gh/masnesral/215/base 2025-07-17T08:34:51.2053200Z * [new branch] gh/masnesral/215/head -> origin/gh/masnesral/215/head 2025-07-17T08:34:51.2056498Z * [new branch] gh/masnesral/215/orig -> origin/gh/masnesral/215/orig 2025-07-17T08:34:51.2061023Z * [new branch] gh/masnesral/216/base -> origin/gh/masnesral/216/base 2025-07-17T08:34:51.2064367Z * [new branch] gh/masnesral/216/head -> origin/gh/masnesral/216/head 2025-07-17T08:34:51.2067722Z * [new branch] gh/masnesral/216/orig -> origin/gh/masnesral/216/orig 2025-07-17T08:34:51.2072334Z * [new branch] gh/masnesral/217/base -> origin/gh/masnesral/217/base 2025-07-17T08:34:51.2075950Z * [new branch] gh/masnesral/217/head -> origin/gh/masnesral/217/head 2025-07-17T08:34:51.2079465Z * [new branch] gh/masnesral/217/orig -> origin/gh/masnesral/217/orig 2025-07-17T08:34:51.2084015Z * [new branch] gh/masnesral/218/base -> origin/gh/masnesral/218/base 2025-07-17T08:34:51.2087370Z * [new branch] gh/masnesral/218/head -> origin/gh/masnesral/218/head 2025-07-17T08:34:51.2090684Z * [new branch] gh/masnesral/218/orig -> origin/gh/masnesral/218/orig 2025-07-17T08:34:51.2095050Z * [new branch] gh/masnesral/219/base -> origin/gh/masnesral/219/base 2025-07-17T08:34:51.2098312Z * [new branch] gh/masnesral/219/head -> origin/gh/masnesral/219/head 2025-07-17T08:34:51.2101580Z * [new branch] gh/masnesral/219/orig -> origin/gh/masnesral/219/orig 2025-07-17T08:34:51.2105910Z * [new branch] gh/masnesral/220/base -> origin/gh/masnesral/220/base 2025-07-17T08:34:51.2109335Z * [new branch] gh/masnesral/220/head -> origin/gh/masnesral/220/head 2025-07-17T08:34:51.2112691Z * [new branch] gh/masnesral/220/orig -> origin/gh/masnesral/220/orig 2025-07-17T08:34:51.2117324Z * [new branch] gh/masnesral/221/base -> origin/gh/masnesral/221/base 2025-07-17T08:34:51.2120614Z * [new branch] gh/masnesral/221/head -> origin/gh/masnesral/221/head 2025-07-17T08:34:51.2124252Z * [new branch] gh/masnesral/221/orig -> origin/gh/masnesral/221/orig 2025-07-17T08:34:51.2128572Z * [new branch] gh/masnesral/222/base -> origin/gh/masnesral/222/base 2025-07-17T08:34:51.2131861Z * [new branch] gh/masnesral/222/head -> origin/gh/masnesral/222/head 2025-07-17T08:34:51.2135130Z * [new branch] gh/masnesral/222/orig -> origin/gh/masnesral/222/orig 2025-07-17T08:34:51.2139504Z * [new branch] gh/masnesral/223/base -> origin/gh/masnesral/223/base 2025-07-17T08:34:51.2142935Z * [new branch] gh/masnesral/223/head -> origin/gh/masnesral/223/head 2025-07-17T08:34:51.2146237Z * [new branch] gh/masnesral/223/orig -> origin/gh/masnesral/223/orig 2025-07-17T08:34:51.2150621Z * [new branch] gh/masnesral/224/base -> origin/gh/masnesral/224/base 2025-07-17T08:34:51.2154231Z * [new branch] gh/masnesral/224/head -> origin/gh/masnesral/224/head 2025-07-17T08:34:51.2157774Z * [new branch] gh/masnesral/224/orig -> origin/gh/masnesral/224/orig 2025-07-17T08:34:51.2162059Z * [new branch] gh/masnesral/225/base -> origin/gh/masnesral/225/base 2025-07-17T08:34:51.2165540Z * [new branch] gh/masnesral/225/head -> origin/gh/masnesral/225/head 2025-07-17T08:34:51.2169016Z * [new branch] gh/masnesral/225/orig -> origin/gh/masnesral/225/orig 2025-07-17T08:34:51.2173643Z * [new branch] gh/masnesral/226/base -> origin/gh/masnesral/226/base 2025-07-17T08:34:51.2177001Z * [new branch] gh/masnesral/226/head -> origin/gh/masnesral/226/head 2025-07-17T08:34:51.2180398Z * [new branch] gh/masnesral/226/orig -> origin/gh/masnesral/226/orig 2025-07-17T08:34:51.2184975Z * [new branch] gh/masnesral/227/base -> origin/gh/masnesral/227/base 2025-07-17T08:34:51.2188354Z * [new branch] gh/masnesral/227/head -> origin/gh/masnesral/227/head 2025-07-17T08:34:51.2191750Z * [new branch] gh/masnesral/227/orig -> origin/gh/masnesral/227/orig 2025-07-17T08:34:51.2196593Z * [new branch] gh/masnesral/228/base -> origin/gh/masnesral/228/base 2025-07-17T08:34:51.2199835Z * [new branch] gh/masnesral/228/head -> origin/gh/masnesral/228/head 2025-07-17T08:34:51.2203232Z * [new branch] gh/masnesral/228/orig -> origin/gh/masnesral/228/orig 2025-07-17T08:34:51.2207830Z * [new branch] gh/masnesral/229/base -> origin/gh/masnesral/229/base 2025-07-17T08:34:51.2211193Z * [new branch] gh/masnesral/229/head -> origin/gh/masnesral/229/head 2025-07-17T08:34:51.2214667Z * [new branch] gh/masnesral/229/orig -> origin/gh/masnesral/229/orig 2025-07-17T08:34:51.2219263Z * [new branch] gh/masnesral/230/base -> origin/gh/masnesral/230/base 2025-07-17T08:34:51.2222673Z * [new branch] gh/masnesral/230/head -> origin/gh/masnesral/230/head 2025-07-17T08:34:51.2226108Z * [new branch] gh/masnesral/230/orig -> origin/gh/masnesral/230/orig 2025-07-17T08:34:51.2230660Z * [new branch] gh/masnesral/34/base -> origin/gh/masnesral/34/base 2025-07-17T08:34:51.2236332Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-07-17T08:34:51.2239830Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-07-17T08:34:51.2244187Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-07-17T08:34:51.2247532Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-07-17T08:34:51.2251948Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-07-17T08:34:51.2255428Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-07-17T08:34:51.2260077Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-07-17T08:34:51.2263424Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-07-17T08:34:51.2267836Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-07-17T08:34:51.2271160Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-07-17T08:34:51.2275796Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-07-17T08:34:51.2279101Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-07-17T08:34:51.2283515Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-07-17T08:34:51.2286856Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-07-17T08:34:51.2292621Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-07-17T08:34:51.2295957Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-07-17T08:34:51.2300477Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-07-17T08:34:51.2303821Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-07-17T08:34:51.2308318Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-07-17T08:34:51.2311625Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-07-17T08:34:51.2316275Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-07-17T08:34:51.2319563Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-07-17T08:34:51.2324065Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-07-17T08:34:51.2327441Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-07-17T08:34:51.2332403Z * [new branch] gh/mikaylagawarecki/313/base -> origin/gh/mikaylagawarecki/313/base 2025-07-17T08:34:51.2335714Z * [new branch] gh/mikaylagawarecki/313/head -> origin/gh/mikaylagawarecki/313/head 2025-07-17T08:34:51.2339087Z * [new branch] gh/mikaylagawarecki/313/orig -> origin/gh/mikaylagawarecki/313/orig 2025-07-17T08:34:51.2343507Z * [new branch] gh/mikaylagawarecki/316/base -> origin/gh/mikaylagawarecki/316/base 2025-07-17T08:34:51.2346835Z * [new branch] gh/mikaylagawarecki/316/head -> origin/gh/mikaylagawarecki/316/head 2025-07-17T08:34:51.2350411Z * [new branch] gh/mikaylagawarecki/316/orig -> origin/gh/mikaylagawarecki/316/orig 2025-07-17T08:34:51.2355217Z * [new branch] gh/mikaylagawarecki/317/base -> origin/gh/mikaylagawarecki/317/base 2025-07-17T08:34:51.2358700Z * [new branch] gh/mikaylagawarecki/317/head -> origin/gh/mikaylagawarecki/317/head 2025-07-17T08:34:51.2362009Z * [new branch] gh/mikaylagawarecki/317/orig -> origin/gh/mikaylagawarecki/317/orig 2025-07-17T08:34:51.2366643Z * [new branch] gh/mikaylagawarecki/318/base -> origin/gh/mikaylagawarecki/318/base 2025-07-17T08:34:51.2369980Z * [new branch] gh/mikaylagawarecki/318/head -> origin/gh/mikaylagawarecki/318/head 2025-07-17T08:34:51.2373318Z * [new branch] gh/mikaylagawarecki/318/orig -> origin/gh/mikaylagawarecki/318/orig 2025-07-17T08:34:51.2378004Z * [new branch] gh/mikaylagawarecki/319/base -> origin/gh/mikaylagawarecki/319/base 2025-07-17T08:34:51.2381313Z * [new branch] gh/mikaylagawarecki/319/head -> origin/gh/mikaylagawarecki/319/head 2025-07-17T08:34:51.2384579Z * [new branch] gh/mikaylagawarecki/319/orig -> origin/gh/mikaylagawarecki/319/orig 2025-07-17T08:34:51.2389424Z * [new branch] gh/mikaylagawarecki/320/base -> origin/gh/mikaylagawarecki/320/base 2025-07-17T08:34:51.2392665Z * [new branch] gh/mikaylagawarecki/320/head -> origin/gh/mikaylagawarecki/320/head 2025-07-17T08:34:51.2396251Z * [new branch] gh/mikaylagawarecki/320/orig -> origin/gh/mikaylagawarecki/320/orig 2025-07-17T08:34:51.2400702Z * [new branch] gh/mikaylagawarecki/321/base -> origin/gh/mikaylagawarecki/321/base 2025-07-17T08:34:51.2404078Z * [new branch] gh/mikaylagawarecki/321/head -> origin/gh/mikaylagawarecki/321/head 2025-07-17T08:34:51.2407505Z * [new branch] gh/mikaylagawarecki/321/orig -> origin/gh/mikaylagawarecki/321/orig 2025-07-17T08:34:51.2412897Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-07-17T08:34:51.2416274Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-07-17T08:34:51.2421935Z * [new branch] gh/muchulee8/62/base -> origin/gh/muchulee8/62/base 2025-07-17T08:34:51.2425336Z * [new branch] gh/muchulee8/62/head -> origin/gh/muchulee8/62/head 2025-07-17T08:34:51.2428773Z * [new branch] gh/muchulee8/62/orig -> origin/gh/muchulee8/62/orig 2025-07-17T08:34:51.2434392Z * [new branch] gh/nautsimon/1/base -> origin/gh/nautsimon/1/base 2025-07-17T08:34:51.2437941Z * [new branch] gh/nautsimon/1/head -> origin/gh/nautsimon/1/head 2025-07-17T08:34:51.2441256Z * [new branch] gh/nautsimon/1/orig -> origin/gh/nautsimon/1/orig 2025-07-17T08:34:51.2445830Z * [new branch] gh/nautsimon/10/base -> origin/gh/nautsimon/10/base 2025-07-17T08:34:51.2449243Z * [new branch] gh/nautsimon/10/head -> origin/gh/nautsimon/10/head 2025-07-17T08:34:51.2452656Z * [new branch] gh/nautsimon/10/orig -> origin/gh/nautsimon/10/orig 2025-07-17T08:34:51.2457041Z * [new branch] gh/nautsimon/11/base -> origin/gh/nautsimon/11/base 2025-07-17T08:34:51.2460425Z * [new branch] gh/nautsimon/11/head -> origin/gh/nautsimon/11/head 2025-07-17T08:34:51.2463754Z * [new branch] gh/nautsimon/11/orig -> origin/gh/nautsimon/11/orig 2025-07-17T08:34:51.2468174Z * [new branch] gh/nautsimon/12/base -> origin/gh/nautsimon/12/base 2025-07-17T08:34:51.2471463Z * [new branch] gh/nautsimon/12/head -> origin/gh/nautsimon/12/head 2025-07-17T08:34:51.2475056Z * [new branch] gh/nautsimon/12/orig -> origin/gh/nautsimon/12/orig 2025-07-17T08:34:51.2479818Z * [new branch] gh/nautsimon/13/base -> origin/gh/nautsimon/13/base 2025-07-17T08:34:51.2483048Z * [new branch] gh/nautsimon/13/head -> origin/gh/nautsimon/13/head 2025-07-17T08:34:51.2486445Z * [new branch] gh/nautsimon/13/orig -> origin/gh/nautsimon/13/orig 2025-07-17T08:34:51.2490902Z * [new branch] gh/nautsimon/14/base -> origin/gh/nautsimon/14/base 2025-07-17T08:34:51.2494269Z * [new branch] gh/nautsimon/14/head -> origin/gh/nautsimon/14/head 2025-07-17T08:34:51.2497667Z * [new branch] gh/nautsimon/14/orig -> origin/gh/nautsimon/14/orig 2025-07-17T08:34:51.2502033Z * [new branch] gh/nautsimon/15/base -> origin/gh/nautsimon/15/base 2025-07-17T08:34:51.2505402Z * [new branch] gh/nautsimon/15/head -> origin/gh/nautsimon/15/head 2025-07-17T08:34:51.2508836Z * [new branch] gh/nautsimon/15/orig -> origin/gh/nautsimon/15/orig 2025-07-17T08:34:51.2513333Z * [new branch] gh/nautsimon/16/base -> origin/gh/nautsimon/16/base 2025-07-17T08:34:51.2516895Z * [new branch] gh/nautsimon/16/head -> origin/gh/nautsimon/16/head 2025-07-17T08:34:51.2520374Z * [new branch] gh/nautsimon/16/orig -> origin/gh/nautsimon/16/orig 2025-07-17T08:34:51.2524560Z * [new branch] gh/nautsimon/17/base -> origin/gh/nautsimon/17/base 2025-07-17T08:34:51.2527915Z * [new branch] gh/nautsimon/17/head -> origin/gh/nautsimon/17/head 2025-07-17T08:34:51.2531323Z * [new branch] gh/nautsimon/17/orig -> origin/gh/nautsimon/17/orig 2025-07-17T08:34:51.2535695Z * [new branch] gh/nautsimon/18/base -> origin/gh/nautsimon/18/base 2025-07-17T08:34:51.2539073Z * [new branch] gh/nautsimon/18/head -> origin/gh/nautsimon/18/head 2025-07-17T08:34:51.2542988Z * [new branch] gh/nautsimon/18/orig -> origin/gh/nautsimon/18/orig 2025-07-17T08:34:51.2547522Z * [new branch] gh/nautsimon/19/base -> origin/gh/nautsimon/19/base 2025-07-17T08:34:51.2550900Z * [new branch] gh/nautsimon/19/head -> origin/gh/nautsimon/19/head 2025-07-17T08:34:51.2554466Z * [new branch] gh/nautsimon/19/orig -> origin/gh/nautsimon/19/orig 2025-07-17T08:34:51.2559097Z * [new branch] gh/nautsimon/2/base -> origin/gh/nautsimon/2/base 2025-07-17T08:34:51.2562490Z * [new branch] gh/nautsimon/2/head -> origin/gh/nautsimon/2/head 2025-07-17T08:34:51.2565811Z * [new branch] gh/nautsimon/2/orig -> origin/gh/nautsimon/2/orig 2025-07-17T08:34:51.2570242Z * [new branch] gh/nautsimon/20/base -> origin/gh/nautsimon/20/base 2025-07-17T08:34:51.2573617Z * [new branch] gh/nautsimon/20/head -> origin/gh/nautsimon/20/head 2025-07-17T08:34:51.2576881Z * [new branch] gh/nautsimon/20/orig -> origin/gh/nautsimon/20/orig 2025-07-17T08:34:51.2581333Z * [new branch] gh/nautsimon/21/base -> origin/gh/nautsimon/21/base 2025-07-17T08:34:51.2584669Z * [new branch] gh/nautsimon/21/head -> origin/gh/nautsimon/21/head 2025-07-17T08:34:51.2588013Z * [new branch] gh/nautsimon/21/orig -> origin/gh/nautsimon/21/orig 2025-07-17T08:34:51.2592449Z * [new branch] gh/nautsimon/3/base -> origin/gh/nautsimon/3/base 2025-07-17T08:34:51.2596105Z * [new branch] gh/nautsimon/3/head -> origin/gh/nautsimon/3/head 2025-07-17T08:34:51.2599545Z * [new branch] gh/nautsimon/3/orig -> origin/gh/nautsimon/3/orig 2025-07-17T08:34:51.2604054Z * [new branch] gh/nautsimon/4/base -> origin/gh/nautsimon/4/base 2025-07-17T08:34:51.2607423Z * [new branch] gh/nautsimon/4/head -> origin/gh/nautsimon/4/head 2025-07-17T08:34:51.2610772Z * [new branch] gh/nautsimon/4/orig -> origin/gh/nautsimon/4/orig 2025-07-17T08:34:51.2615279Z * [new branch] gh/nautsimon/5/base -> origin/gh/nautsimon/5/base 2025-07-17T08:34:51.2618545Z * [new branch] gh/nautsimon/5/head -> origin/gh/nautsimon/5/head 2025-07-17T08:34:51.2621896Z * [new branch] gh/nautsimon/5/orig -> origin/gh/nautsimon/5/orig 2025-07-17T08:34:51.2626330Z * [new branch] gh/nautsimon/6/base -> origin/gh/nautsimon/6/base 2025-07-17T08:34:51.2629639Z * [new branch] gh/nautsimon/6/head -> origin/gh/nautsimon/6/head 2025-07-17T08:34:51.2632983Z * [new branch] gh/nautsimon/6/orig -> origin/gh/nautsimon/6/orig 2025-07-17T08:34:51.2638181Z * [new branch] gh/nautsimon/7/base -> origin/gh/nautsimon/7/base 2025-07-17T08:34:51.2641527Z * [new branch] gh/nautsimon/7/head -> origin/gh/nautsimon/7/head 2025-07-17T08:34:51.2644962Z * [new branch] gh/nautsimon/7/orig -> origin/gh/nautsimon/7/orig 2025-07-17T08:34:51.2649398Z * [new branch] gh/nautsimon/8/base -> origin/gh/nautsimon/8/base 2025-07-17T08:34:51.2652749Z * [new branch] gh/nautsimon/8/head -> origin/gh/nautsimon/8/head 2025-07-17T08:34:51.2656283Z * [new branch] gh/nautsimon/8/orig -> origin/gh/nautsimon/8/orig 2025-07-17T08:34:51.2660524Z * [new branch] gh/nautsimon/9/base -> origin/gh/nautsimon/9/base 2025-07-17T08:34:51.2663725Z * [new branch] gh/nautsimon/9/head -> origin/gh/nautsimon/9/head 2025-07-17T08:34:51.2667063Z * [new branch] gh/nautsimon/9/orig -> origin/gh/nautsimon/9/orig 2025-07-17T08:34:51.2672493Z * [new branch] gh/oulgen/35/base -> origin/gh/oulgen/35/base 2025-07-17T08:34:51.2676090Z * [new branch] gh/oulgen/35/head -> origin/gh/oulgen/35/head 2025-07-17T08:34:51.2679479Z * [new branch] gh/oulgen/35/orig -> origin/gh/oulgen/35/orig 2025-07-17T08:34:51.2683970Z * [new branch] gh/oulgen/38/base -> origin/gh/oulgen/38/base 2025-07-17T08:34:51.2687498Z * [new branch] gh/oulgen/38/head -> origin/gh/oulgen/38/head 2025-07-17T08:34:51.2690839Z * [new branch] gh/oulgen/38/orig -> origin/gh/oulgen/38/orig 2025-07-17T08:34:51.2695402Z * [new branch] gh/oulgen/42/base -> origin/gh/oulgen/42/base 2025-07-17T08:34:51.2698704Z * [new branch] gh/oulgen/42/head -> origin/gh/oulgen/42/head 2025-07-17T08:34:51.2702053Z * [new branch] gh/oulgen/42/orig -> origin/gh/oulgen/42/orig 2025-07-17T08:34:51.2706659Z * [new branch] gh/oulgen/43/base -> origin/gh/oulgen/43/base 2025-07-17T08:34:51.2709975Z * [new branch] gh/oulgen/43/head -> origin/gh/oulgen/43/head 2025-07-17T08:34:51.2713320Z * [new branch] gh/oulgen/43/orig -> origin/gh/oulgen/43/orig 2025-07-17T08:34:51.2718109Z * [new branch] gh/oulgen/44/base -> origin/gh/oulgen/44/base 2025-07-17T08:34:51.2721369Z * [new branch] gh/oulgen/44/head -> origin/gh/oulgen/44/head 2025-07-17T08:34:51.2724659Z * [new branch] gh/oulgen/44/orig -> origin/gh/oulgen/44/orig 2025-07-17T08:34:51.2730377Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-07-17T08:34:51.2734040Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-07-17T08:34:51.2737490Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-07-17T08:34:51.2742043Z * [new branch] gh/pearu/113/base -> origin/gh/pearu/113/base 2025-07-17T08:34:51.2745363Z * [new branch] gh/pearu/113/head -> origin/gh/pearu/113/head 2025-07-17T08:34:51.2748677Z * [new branch] gh/pearu/113/orig -> origin/gh/pearu/113/orig 2025-07-17T08:34:51.2753643Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-07-17T08:34:51.2757350Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-07-17T08:34:51.2760607Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-07-17T08:34:51.2765295Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-07-17T08:34:51.2768642Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-07-17T08:34:51.2772033Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-07-17T08:34:51.2777515Z * [new branch] gh/qqaatw/29/base -> origin/gh/qqaatw/29/base 2025-07-17T08:34:51.2780799Z * [new branch] gh/qqaatw/29/head -> origin/gh/qqaatw/29/head 2025-07-17T08:34:51.2784162Z * [new branch] gh/qqaatw/29/orig -> origin/gh/qqaatw/29/orig 2025-07-17T08:34:51.2788719Z * [new branch] gh/raymo/cleanup-dynamo-logging -> origin/gh/raymo/cleanup-dynamo-logging 2025-07-17T08:34:51.2792173Z * [new branch] gh/raymo/cuda-error-exp -> origin/gh/raymo/cuda-error-exp 2025-07-17T08:34:51.2795400Z * [new branch] gh/raymo/refresh-script -> origin/gh/raymo/refresh-script 2025-07-17T08:34:51.2800946Z * [new branch] gh/rec/133/base -> origin/gh/rec/133/base 2025-07-17T08:34:51.2804174Z * [new branch] gh/rec/133/head -> origin/gh/rec/133/head 2025-07-17T08:34:51.2807476Z * [new branch] gh/rec/133/orig -> origin/gh/rec/133/orig 2025-07-17T08:34:51.2811997Z * [new branch] gh/rec/137/base -> origin/gh/rec/137/base 2025-07-17T08:34:51.2815386Z * [new branch] gh/rec/137/head -> origin/gh/rec/137/head 2025-07-17T08:34:51.2818812Z * [new branch] gh/rec/137/orig -> origin/gh/rec/137/orig 2025-07-17T08:34:51.2823334Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-07-17T08:34:51.2826657Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-07-17T08:34:51.2831195Z * [new branch] gh/rec/142/base -> origin/gh/rec/142/base 2025-07-17T08:34:51.2834802Z * [new branch] gh/rec/142/head -> origin/gh/rec/142/head 2025-07-17T08:34:51.2838241Z * [new branch] gh/rec/142/orig -> origin/gh/rec/142/orig 2025-07-17T08:34:51.2842645Z * [new branch] gh/rec/146/base -> origin/gh/rec/146/base 2025-07-17T08:34:51.2845991Z * [new branch] gh/rec/146/head -> origin/gh/rec/146/head 2025-07-17T08:34:51.2849340Z * [new branch] gh/rec/146/orig -> origin/gh/rec/146/orig 2025-07-17T08:34:51.2853950Z * [new branch] gh/rec/148/base -> origin/gh/rec/148/base 2025-07-17T08:34:51.2857467Z * [new branch] gh/rec/148/head -> origin/gh/rec/148/head 2025-07-17T08:34:51.2860798Z * [new branch] gh/rec/148/orig -> origin/gh/rec/148/orig 2025-07-17T08:34:51.2865434Z * [new branch] gh/rec/150/base -> origin/gh/rec/150/base 2025-07-17T08:34:51.2868812Z * [new branch] gh/rec/150/head -> origin/gh/rec/150/head 2025-07-17T08:34:51.2872125Z * [new branch] gh/rec/150/orig -> origin/gh/rec/150/orig 2025-07-17T08:34:51.2884065Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-07-17T08:34:51.2887323Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-07-17T08:34:51.2890645Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-07-17T08:34:51.2895177Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-07-17T08:34:51.2898497Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-07-17T08:34:51.2901911Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-07-17T08:34:51.2906445Z * [new branch] gh/rec/156/base -> origin/gh/rec/156/base 2025-07-17T08:34:51.2909791Z * [new branch] gh/rec/156/head -> origin/gh/rec/156/head 2025-07-17T08:34:51.2913068Z * [new branch] gh/rec/156/orig -> origin/gh/rec/156/orig 2025-07-17T08:34:51.2917973Z * [new branch] gh/rec/157/base -> origin/gh/rec/157/base 2025-07-17T08:34:51.2921279Z * [new branch] gh/rec/157/head -> origin/gh/rec/157/head 2025-07-17T08:34:51.2924626Z * [new branch] gh/rec/157/orig -> origin/gh/rec/157/orig 2025-07-17T08:34:51.2930180Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-07-17T08:34:51.2933752Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-07-17T08:34:51.2937131Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-07-17T08:34:51.2941847Z * [new branch] gh/ruisizhang123/2/base -> origin/gh/ruisizhang123/2/base 2025-07-17T08:34:51.2944978Z * [new branch] gh/ruisizhang123/2/head -> origin/gh/ruisizhang123/2/head 2025-07-17T08:34:51.2948284Z * [new branch] gh/ruisizhang123/2/orig -> origin/gh/ruisizhang123/2/orig 2025-07-17T08:34:51.2952918Z * [new branch] gh/ruisizhang123/3/base -> origin/gh/ruisizhang123/3/base 2025-07-17T08:34:51.2956461Z * [new branch] gh/ruisizhang123/3/head -> origin/gh/ruisizhang123/3/head 2025-07-17T08:34:51.2959881Z * [new branch] gh/ruisizhang123/3/orig -> origin/gh/ruisizhang123/3/orig 2025-07-17T08:34:51.2964335Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-07-17T08:34:51.2967711Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-07-17T08:34:51.2971039Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-07-17T08:34:51.2975480Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-07-17T08:34:51.2978748Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-07-17T08:34:51.2982072Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-07-17T08:34:51.2986489Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-07-17T08:34:51.2989815Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-07-17T08:34:51.2993199Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-07-17T08:34:51.2998882Z * [new branch] gh/sarckk/2/base -> origin/gh/sarckk/2/base 2025-07-17T08:34:51.3002174Z * [new branch] gh/sarckk/2/head -> origin/gh/sarckk/2/head 2025-07-17T08:34:51.3005527Z * [new branch] gh/sarckk/2/orig -> origin/gh/sarckk/2/orig 2025-07-17T08:34:51.3011115Z * [new branch] gh/seemethere/14/base -> origin/gh/seemethere/14/base 2025-07-17T08:34:51.3014501Z * [new branch] gh/seemethere/14/head -> origin/gh/seemethere/14/head 2025-07-17T08:34:51.3017857Z * [new branch] gh/seemethere/14/orig -> origin/gh/seemethere/14/orig 2025-07-17T08:34:51.3022365Z * [new branch] gh/seemethere/23/base -> origin/gh/seemethere/23/base 2025-07-17T08:34:51.3025729Z * [new branch] gh/seemethere/23/head -> origin/gh/seemethere/23/head 2025-07-17T08:34:51.3029059Z * [new branch] gh/seemethere/23/orig -> origin/gh/seemethere/23/orig 2025-07-17T08:34:51.3033651Z * [new branch] gh/seemethere/24/base -> origin/gh/seemethere/24/base 2025-07-17T08:34:51.3037272Z * [new branch] gh/seemethere/24/head -> origin/gh/seemethere/24/head 2025-07-17T08:34:51.3040641Z * [new branch] gh/seemethere/24/orig -> origin/gh/seemethere/24/orig 2025-07-17T08:34:51.3045181Z * [new branch] gh/seemethere/30/base -> origin/gh/seemethere/30/base 2025-07-17T08:34:51.3048500Z * [new branch] gh/seemethere/30/head -> origin/gh/seemethere/30/head 2025-07-17T08:34:51.3051920Z * [new branch] gh/seemethere/30/orig -> origin/gh/seemethere/30/orig 2025-07-17T08:34:51.3056456Z * [new branch] gh/seemethere/32/base -> origin/gh/seemethere/32/base 2025-07-17T08:34:51.3059770Z * [new branch] gh/seemethere/32/head -> origin/gh/seemethere/32/head 2025-07-17T08:34:51.3063140Z * [new branch] gh/seemethere/32/orig -> origin/gh/seemethere/32/orig 2025-07-17T08:34:51.3067703Z * [new branch] gh/seemethere/33/base -> origin/gh/seemethere/33/base 2025-07-17T08:34:51.3071182Z * [new branch] gh/seemethere/33/head -> origin/gh/seemethere/33/head 2025-07-17T08:34:51.3074485Z * [new branch] gh/seemethere/33/orig -> origin/gh/seemethere/33/orig 2025-07-17T08:34:51.3079197Z * [new branch] gh/seemethere/34/base -> origin/gh/seemethere/34/base 2025-07-17T08:34:51.3082450Z * [new branch] gh/seemethere/34/head -> origin/gh/seemethere/34/head 2025-07-17T08:34:51.3085792Z * [new branch] gh/seemethere/34/orig -> origin/gh/seemethere/34/orig 2025-07-17T08:34:51.3090526Z * [new branch] gh/seemethere/35/base -> origin/gh/seemethere/35/base 2025-07-17T08:34:51.3093975Z * [new branch] gh/seemethere/35/head -> origin/gh/seemethere/35/head 2025-07-17T08:34:51.3097313Z * [new branch] gh/seemethere/35/orig -> origin/gh/seemethere/35/orig 2025-07-17T08:34:51.3101813Z * [new branch] gh/seemethere/36/base -> origin/gh/seemethere/36/base 2025-07-17T08:34:51.3105100Z * [new branch] gh/seemethere/36/head -> origin/gh/seemethere/36/head 2025-07-17T08:34:51.3108495Z * [new branch] gh/seemethere/36/orig -> origin/gh/seemethere/36/orig 2025-07-17T08:34:51.3112928Z * [new branch] gh/seemethere/37/base -> origin/gh/seemethere/37/base 2025-07-17T08:34:51.3117895Z * [new branch] gh/seemethere/37/head -> origin/gh/seemethere/37/head 2025-07-17T08:34:51.3121170Z * [new branch] gh/seemethere/37/orig -> origin/gh/seemethere/37/orig 2025-07-17T08:34:51.3125918Z * [new branch] gh/seemethere/38/base -> origin/gh/seemethere/38/base 2025-07-17T08:34:51.3129451Z * [new branch] gh/seemethere/38/head -> origin/gh/seemethere/38/head 2025-07-17T08:34:51.3132931Z * [new branch] gh/seemethere/38/orig -> origin/gh/seemethere/38/orig 2025-07-17T08:34:51.3137496Z * [new branch] gh/seemethere/39/base -> origin/gh/seemethere/39/base 2025-07-17T08:34:51.3140845Z * [new branch] gh/seemethere/39/head -> origin/gh/seemethere/39/head 2025-07-17T08:34:51.3144173Z * [new branch] gh/seemethere/39/orig -> origin/gh/seemethere/39/orig 2025-07-17T08:34:51.3148695Z * [new branch] gh/seemethere/40/base -> origin/gh/seemethere/40/base 2025-07-17T08:34:51.3152031Z * [new branch] gh/seemethere/40/head -> origin/gh/seemethere/40/head 2025-07-17T08:34:51.3155605Z * [new branch] gh/seemethere/40/orig -> origin/gh/seemethere/40/orig 2025-07-17T08:34:51.3160213Z * [new branch] gh/seemethere/41/base -> origin/gh/seemethere/41/base 2025-07-17T08:34:51.3163428Z * [new branch] gh/seemethere/41/head -> origin/gh/seemethere/41/head 2025-07-17T08:34:51.3166791Z * [new branch] gh/seemethere/41/orig -> origin/gh/seemethere/41/orig 2025-07-17T08:34:51.3171356Z * [new branch] gh/seemethere/42/base -> origin/gh/seemethere/42/base 2025-07-17T08:34:51.3174668Z * [new branch] gh/seemethere/42/head -> origin/gh/seemethere/42/head 2025-07-17T08:34:51.3178029Z * [new branch] gh/seemethere/42/orig -> origin/gh/seemethere/42/orig 2025-07-17T08:34:51.3182798Z * [new branch] gh/seemethere/43/base -> origin/gh/seemethere/43/base 2025-07-17T08:34:51.3186108Z * [new branch] gh/seemethere/43/head -> origin/gh/seemethere/43/head 2025-07-17T08:34:51.3189514Z * [new branch] gh/seemethere/43/orig -> origin/gh/seemethere/43/orig 2025-07-17T08:34:51.3194024Z * [new branch] gh/seemethere/44/base -> origin/gh/seemethere/44/base 2025-07-17T08:34:51.3197457Z * [new branch] gh/seemethere/44/head -> origin/gh/seemethere/44/head 2025-07-17T08:34:51.3200765Z * [new branch] gh/seemethere/44/orig -> origin/gh/seemethere/44/orig 2025-07-17T08:34:51.3205406Z * [new branch] gh/seemethere/45/base -> origin/gh/seemethere/45/base 2025-07-17T08:34:51.3208578Z * [new branch] gh/seemethere/45/head -> origin/gh/seemethere/45/head 2025-07-17T08:34:51.3211932Z * [new branch] gh/seemethere/45/orig -> origin/gh/seemethere/45/orig 2025-07-17T08:34:51.3216435Z * [new branch] gh/seemethere/46/base -> origin/gh/seemethere/46/base 2025-07-17T08:34:51.3219710Z * [new branch] gh/seemethere/46/head -> origin/gh/seemethere/46/head 2025-07-17T08:34:51.3223086Z * [new branch] gh/seemethere/46/orig -> origin/gh/seemethere/46/orig 2025-07-17T08:34:51.3227713Z * [new branch] gh/seemethere/47/base -> origin/gh/seemethere/47/base 2025-07-17T08:34:51.3231079Z * [new branch] gh/seemethere/47/head -> origin/gh/seemethere/47/head 2025-07-17T08:34:51.3234590Z * [new branch] gh/seemethere/47/orig -> origin/gh/seemethere/47/orig 2025-07-17T08:34:51.3239218Z * [new branch] gh/seemethere/48/base -> origin/gh/seemethere/48/base 2025-07-17T08:34:51.3242466Z * [new branch] gh/seemethere/48/head -> origin/gh/seemethere/48/head 2025-07-17T08:34:51.3245850Z * [new branch] gh/seemethere/48/orig -> origin/gh/seemethere/48/orig 2025-07-17T08:34:51.3252008Z * [new branch] gh/seemethere/7/base -> origin/gh/seemethere/7/base 2025-07-17T08:34:51.3253798Z * [new branch] gh/seemethere/7/head -> origin/gh/seemethere/7/head 2025-07-17T08:34:51.3257123Z * [new branch] gh/seemethere/7/orig -> origin/gh/seemethere/7/orig 2025-07-17T08:34:51.3262726Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-07-17T08:34:51.3266244Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-07-17T08:34:51.3269765Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-07-17T08:34:51.3274898Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-07-17T08:34:51.3278500Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-07-17T08:34:51.3281804Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-07-17T08:34:51.3286791Z * [new branch] gh/shunting314/210/base -> origin/gh/shunting314/210/base 2025-07-17T08:34:51.3290491Z * [new branch] gh/shunting314/210/head -> origin/gh/shunting314/210/head 2025-07-17T08:34:51.3293882Z * [new branch] gh/shunting314/210/orig -> origin/gh/shunting314/210/orig 2025-07-17T08:34:51.3298433Z * [new branch] gh/shunting314/211/base -> origin/gh/shunting314/211/base 2025-07-17T08:34:51.3301749Z * [new branch] gh/shunting314/211/head -> origin/gh/shunting314/211/head 2025-07-17T08:34:51.3305085Z * [new branch] gh/shunting314/211/orig -> origin/gh/shunting314/211/orig 2025-07-17T08:34:51.3309534Z * [new branch] gh/shunting314/212/base -> origin/gh/shunting314/212/base 2025-07-17T08:34:51.3312847Z * [new branch] gh/shunting314/212/head -> origin/gh/shunting314/212/head 2025-07-17T08:34:51.3316706Z * [new branch] gh/shunting314/212/orig -> origin/gh/shunting314/212/orig 2025-07-17T08:34:51.3321727Z * [new branch] gh/shunting314/213/base -> origin/gh/shunting314/213/base 2025-07-17T08:34:51.3325172Z * [new branch] gh/shunting314/213/head -> origin/gh/shunting314/213/head 2025-07-17T08:34:51.3328576Z * [new branch] gh/shunting314/213/orig -> origin/gh/shunting314/213/orig 2025-07-17T08:34:51.3334042Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-07-17T08:34:51.3337598Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-07-17T08:34:51.3341835Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-07-17T08:34:51.3345109Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-07-17T08:34:51.3349499Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-07-17T08:34:51.3352803Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-07-17T08:34:51.3358045Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-07-17T08:34:51.3361408Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-07-17T08:34:51.3366825Z * [new branch] gh/sinhaanhsul/1/base -> origin/gh/sinhaanhsul/1/base 2025-07-17T08:34:51.3370251Z * [new branch] gh/sinhaanhsul/1/head -> origin/gh/sinhaanhsul/1/head 2025-07-17T08:34:51.3375891Z * [new branch] gh/skarjala/1/base -> origin/gh/skarjala/1/base 2025-07-17T08:34:51.3379236Z * [new branch] gh/skarjala/1/head -> origin/gh/skarjala/1/head 2025-07-17T08:34:51.3382600Z * [new branch] gh/skarjala/1/orig -> origin/gh/skarjala/1/orig 2025-07-17T08:34:51.3387242Z * [new branch] gh/skarjala/10/base -> origin/gh/skarjala/10/base 2025-07-17T08:34:51.3390609Z * [new branch] gh/skarjala/10/head -> origin/gh/skarjala/10/head 2025-07-17T08:34:51.3394060Z * [new branch] gh/skarjala/10/orig -> origin/gh/skarjala/10/orig 2025-07-17T08:34:51.3398755Z * [new branch] gh/skarjala/11/base -> origin/gh/skarjala/11/base 2025-07-17T08:34:51.3402192Z * [new branch] gh/skarjala/11/head -> origin/gh/skarjala/11/head 2025-07-17T08:34:51.3405550Z * [new branch] gh/skarjala/11/orig -> origin/gh/skarjala/11/orig 2025-07-17T08:34:51.3410057Z * [new branch] gh/skarjala/12/base -> origin/gh/skarjala/12/base 2025-07-17T08:34:51.3413614Z * [new branch] gh/skarjala/12/head -> origin/gh/skarjala/12/head 2025-07-17T08:34:51.3416914Z * [new branch] gh/skarjala/12/orig -> origin/gh/skarjala/12/orig 2025-07-17T08:34:51.3421368Z * [new branch] gh/skarjala/2/base -> origin/gh/skarjala/2/base 2025-07-17T08:34:51.3424673Z * [new branch] gh/skarjala/2/head -> origin/gh/skarjala/2/head 2025-07-17T08:34:51.3428053Z * [new branch] gh/skarjala/2/orig -> origin/gh/skarjala/2/orig 2025-07-17T08:34:51.3432620Z * [new branch] gh/skarjala/3/base -> origin/gh/skarjala/3/base 2025-07-17T08:34:51.3436167Z * [new branch] gh/skarjala/3/head -> origin/gh/skarjala/3/head 2025-07-17T08:34:51.3439538Z * [new branch] gh/skarjala/3/orig -> origin/gh/skarjala/3/orig 2025-07-17T08:34:51.3444067Z * [new branch] gh/skarjala/4/base -> origin/gh/skarjala/4/base 2025-07-17T08:34:51.3447460Z * [new branch] gh/skarjala/4/head -> origin/gh/skarjala/4/head 2025-07-17T08:34:51.3450830Z * [new branch] gh/skarjala/4/orig -> origin/gh/skarjala/4/orig 2025-07-17T08:34:51.3455518Z * [new branch] gh/skarjala/5/base -> origin/gh/skarjala/5/base 2025-07-17T08:34:51.3458970Z * [new branch] gh/skarjala/5/head -> origin/gh/skarjala/5/head 2025-07-17T08:34:51.3462275Z * [new branch] gh/skarjala/5/orig -> origin/gh/skarjala/5/orig 2025-07-17T08:34:51.3466809Z * [new branch] gh/skarjala/6/base -> origin/gh/skarjala/6/base 2025-07-17T08:34:51.3470130Z * [new branch] gh/skarjala/6/head -> origin/gh/skarjala/6/head 2025-07-17T08:34:51.3473469Z * [new branch] gh/skarjala/6/orig -> origin/gh/skarjala/6/orig 2025-07-17T08:34:51.3478317Z * [new branch] gh/skarjala/7/base -> origin/gh/skarjala/7/base 2025-07-17T08:34:51.3481425Z * [new branch] gh/skarjala/7/head -> origin/gh/skarjala/7/head 2025-07-17T08:34:51.3484779Z * [new branch] gh/skarjala/7/orig -> origin/gh/skarjala/7/orig 2025-07-17T08:34:51.3489281Z * [new branch] gh/skarjala/8/base -> origin/gh/skarjala/8/base 2025-07-17T08:34:51.3492708Z * [new branch] gh/skarjala/8/head -> origin/gh/skarjala/8/head 2025-07-17T08:34:51.3496050Z * [new branch] gh/skarjala/8/orig -> origin/gh/skarjala/8/orig 2025-07-17T08:34:51.3500550Z * [new branch] gh/skarjala/9/base -> origin/gh/skarjala/9/base 2025-07-17T08:34:51.3503892Z * [new branch] gh/skarjala/9/head -> origin/gh/skarjala/9/head 2025-07-17T08:34:51.3507199Z * [new branch] gh/skarjala/9/orig -> origin/gh/skarjala/9/orig 2025-07-17T08:34:51.3512961Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-07-17T08:34:51.3516479Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-07-17T08:34:51.3519895Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-07-17T08:34:51.3524467Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-07-17T08:34:51.3527848Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-07-17T08:34:51.3531267Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-07-17T08:34:51.3536205Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-07-17T08:34:51.3539561Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-07-17T08:34:51.3543014Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-07-17T08:34:51.3547620Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-07-17T08:34:51.3551005Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-07-17T08:34:51.3554442Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-07-17T08:34:51.3559023Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-07-17T08:34:51.3562392Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-07-17T08:34:51.3565754Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-07-17T08:34:51.3570291Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-07-17T08:34:51.3573639Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-07-17T08:34:51.3576991Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-07-17T08:34:51.3581698Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-07-17T08:34:51.3585402Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-07-17T08:34:51.3588784Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-07-17T08:34:51.3593527Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-07-17T08:34:51.3599541Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-07-17T08:34:51.3603118Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-07-17T08:34:51.3607731Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-07-17T08:34:51.3611109Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-07-17T08:34:51.3614695Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-07-17T08:34:51.3619157Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-07-17T08:34:51.3622471Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-07-17T08:34:51.3625823Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-07-17T08:34:51.3630494Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-07-17T08:34:51.3633839Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-07-17T08:34:51.3637432Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-07-17T08:34:51.3642087Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-07-17T08:34:51.3645371Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-07-17T08:34:51.3648798Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-07-17T08:34:51.3662813Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-07-17T08:34:51.3663259Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-07-17T08:34:51.3663593Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-07-17T08:34:51.3664967Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-07-17T08:34:51.3668235Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-07-17T08:34:51.3671538Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-07-17T08:34:51.3676406Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-07-17T08:34:51.3679854Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-07-17T08:34:51.3683157Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-07-17T08:34:51.3687824Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-07-17T08:34:51.3691339Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-07-17T08:34:51.3694913Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-07-17T08:34:51.3699720Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-07-17T08:34:51.3703292Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-07-17T08:34:51.3706924Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-07-17T08:34:51.3711870Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-07-17T08:34:51.3715554Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-07-17T08:34:51.3718890Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-07-17T08:34:51.3723444Z * [new branch] gh/soulitzer/362/base -> origin/gh/soulitzer/362/base 2025-07-17T08:34:51.3726924Z * [new branch] gh/soulitzer/362/head -> origin/gh/soulitzer/362/head 2025-07-17T08:34:51.3730297Z * [new branch] gh/soulitzer/362/orig -> origin/gh/soulitzer/362/orig 2025-07-17T08:34:51.3734915Z * [new branch] gh/soulitzer/371/base -> origin/gh/soulitzer/371/base 2025-07-17T08:34:51.3738244Z * [new branch] gh/soulitzer/371/head -> origin/gh/soulitzer/371/head 2025-07-17T08:34:51.3741714Z * [new branch] gh/soulitzer/371/orig -> origin/gh/soulitzer/371/orig 2025-07-17T08:34:51.3746532Z * [new branch] gh/soulitzer/372/base -> origin/gh/soulitzer/372/base 2025-07-17T08:34:51.3749688Z * [new branch] gh/soulitzer/372/head -> origin/gh/soulitzer/372/head 2025-07-17T08:34:51.3753054Z * [new branch] gh/soulitzer/372/orig -> origin/gh/soulitzer/372/orig 2025-07-17T08:34:51.3757878Z * [new branch] gh/soulitzer/373/base -> origin/gh/soulitzer/373/base 2025-07-17T08:34:51.3761234Z * [new branch] gh/soulitzer/373/head -> origin/gh/soulitzer/373/head 2025-07-17T08:34:51.3764636Z * [new branch] gh/soulitzer/373/orig -> origin/gh/soulitzer/373/orig 2025-07-17T08:34:51.3769239Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-07-17T08:34:51.3772559Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-07-17T08:34:51.3775950Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-07-17T08:34:51.3781486Z * [new branch] gh/suo/619/base -> origin/gh/suo/619/base 2025-07-17T08:34:51.3787089Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-07-17T08:34:51.3791677Z * [new branch] gh/swolchok/758/base -> origin/gh/swolchok/758/base 2025-07-17T08:34:51.3795177Z * [new branch] gh/swolchok/758/head -> origin/gh/swolchok/758/head 2025-07-17T08:34:51.3798594Z * [new branch] gh/swolchok/758/orig -> origin/gh/swolchok/758/orig 2025-07-17T08:34:51.3803413Z * [new branch] gh/swolchok/767/base -> origin/gh/swolchok/767/base 2025-07-17T08:34:51.3807012Z * [new branch] gh/swolchok/767/head -> origin/gh/swolchok/767/head 2025-07-17T08:34:51.3810644Z * [new branch] gh/swolchok/767/orig -> origin/gh/swolchok/767/orig 2025-07-17T08:34:51.3815256Z * [new branch] gh/swolchok/768/base -> origin/gh/swolchok/768/base 2025-07-17T08:34:51.3818710Z * [new branch] gh/swolchok/768/head -> origin/gh/swolchok/768/head 2025-07-17T08:34:51.3822137Z * [new branch] gh/swolchok/768/orig -> origin/gh/swolchok/768/orig 2025-07-17T08:34:51.3826989Z * [new branch] gh/swolchok/769/base -> origin/gh/swolchok/769/base 2025-07-17T08:34:51.3830376Z * [new branch] gh/swolchok/769/head -> origin/gh/swolchok/769/head 2025-07-17T08:34:51.3833968Z * [new branch] gh/swolchok/769/orig -> origin/gh/swolchok/769/orig 2025-07-17T08:34:51.3838923Z * [new branch] gh/swolchok/771/base -> origin/gh/swolchok/771/base 2025-07-17T08:34:51.3842274Z * [new branch] gh/swolchok/771/head -> origin/gh/swolchok/771/head 2025-07-17T08:34:51.3845782Z * [new branch] gh/swolchok/771/orig -> origin/gh/swolchok/771/orig 2025-07-17T08:34:51.3850418Z * [new branch] gh/swolchok/772/base -> origin/gh/swolchok/772/base 2025-07-17T08:34:51.3853841Z * [new branch] gh/swolchok/772/head -> origin/gh/swolchok/772/head 2025-07-17T08:34:51.3857292Z * [new branch] gh/swolchok/772/orig -> origin/gh/swolchok/772/orig 2025-07-17T08:34:51.3862066Z * [new branch] gh/swolchok/773/base -> origin/gh/swolchok/773/base 2025-07-17T08:34:51.3865529Z * [new branch] gh/swolchok/773/head -> origin/gh/swolchok/773/head 2025-07-17T08:34:51.3868970Z * [new branch] gh/swolchok/773/orig -> origin/gh/swolchok/773/orig 2025-07-17T08:34:51.3873602Z * [new branch] gh/swolchok/775/base -> origin/gh/swolchok/775/base 2025-07-17T08:34:51.3877235Z * [new branch] gh/swolchok/775/head -> origin/gh/swolchok/775/head 2025-07-17T08:34:51.3880581Z * [new branch] gh/swolchok/775/orig -> origin/gh/swolchok/775/orig 2025-07-17T08:34:51.3885273Z * [new branch] gh/swolchok/776/base -> origin/gh/swolchok/776/base 2025-07-17T08:34:51.3888641Z * [new branch] gh/swolchok/776/head -> origin/gh/swolchok/776/head 2025-07-17T08:34:51.3892061Z * [new branch] gh/swolchok/776/orig -> origin/gh/swolchok/776/orig 2025-07-17T08:34:51.3896704Z * [new branch] gh/swolchok/777/base -> origin/gh/swolchok/777/base 2025-07-17T08:34:51.3900057Z * [new branch] gh/swolchok/777/head -> origin/gh/swolchok/777/head 2025-07-17T08:34:51.3903708Z * [new branch] gh/swolchok/777/orig -> origin/gh/swolchok/777/orig 2025-07-17T08:34:51.3908405Z * [new branch] gh/swolchok/778/base -> origin/gh/swolchok/778/base 2025-07-17T08:34:51.3911767Z * [new branch] gh/swolchok/778/head -> origin/gh/swolchok/778/head 2025-07-17T08:34:51.3915332Z * [new branch] gh/swolchok/778/orig -> origin/gh/swolchok/778/orig 2025-07-17T08:34:51.3920125Z * [new branch] gh/swolchok/779/base -> origin/gh/swolchok/779/base 2025-07-17T08:34:51.3923446Z * [new branch] gh/swolchok/779/head -> origin/gh/swolchok/779/head 2025-07-17T08:34:51.3926828Z * [new branch] gh/swolchok/779/orig -> origin/gh/swolchok/779/orig 2025-07-17T08:34:51.3931524Z * [new branch] gh/swolchok/780/base -> origin/gh/swolchok/780/base 2025-07-17T08:34:51.3934929Z * [new branch] gh/swolchok/780/head -> origin/gh/swolchok/780/head 2025-07-17T08:34:51.3938275Z * [new branch] gh/swolchok/780/orig -> origin/gh/swolchok/780/orig 2025-07-17T08:34:51.3942859Z * [new branch] gh/swolchok/781/base -> origin/gh/swolchok/781/base 2025-07-17T08:34:51.3946229Z * [new branch] gh/swolchok/781/head -> origin/gh/swolchok/781/head 2025-07-17T08:34:51.3949584Z * [new branch] gh/swolchok/781/orig -> origin/gh/swolchok/781/orig 2025-07-17T08:34:51.3954404Z * [new branch] gh/swolchok/782/base -> origin/gh/swolchok/782/base 2025-07-17T08:34:51.3957837Z * [new branch] gh/swolchok/782/head -> origin/gh/swolchok/782/head 2025-07-17T08:34:51.3961116Z * [new branch] gh/swolchok/782/orig -> origin/gh/swolchok/782/orig 2025-07-17T08:34:51.3965811Z * [new branch] gh/swolchok/783/base -> origin/gh/swolchok/783/base 2025-07-17T08:34:51.3969156Z * [new branch] gh/swolchok/783/head -> origin/gh/swolchok/783/head 2025-07-17T08:34:51.3972525Z * [new branch] gh/swolchok/783/orig -> origin/gh/swolchok/783/orig 2025-07-17T08:34:51.3977109Z * [new branch] gh/swolchok/784/base -> origin/gh/swolchok/784/base 2025-07-17T08:34:51.3980546Z * [new branch] gh/swolchok/784/head -> origin/gh/swolchok/784/head 2025-07-17T08:34:51.3984025Z * [new branch] gh/swolchok/784/orig -> origin/gh/swolchok/784/orig 2025-07-17T08:34:51.3988553Z * [new branch] gh/swolchok/785/base -> origin/gh/swolchok/785/base 2025-07-17T08:34:51.3995024Z * [new branch] gh/swolchok/785/head -> origin/gh/swolchok/785/head 2025-07-17T08:34:51.3998437Z * [new branch] gh/swolchok/785/orig -> origin/gh/swolchok/785/orig 2025-07-17T08:34:51.4004048Z * [new branch] gh/syed-ahmed/2/base -> origin/gh/syed-ahmed/2/base 2025-07-17T08:34:51.4007398Z * [new branch] gh/syed-ahmed/2/head -> origin/gh/syed-ahmed/2/head 2025-07-17T08:34:51.4010837Z * [new branch] gh/syed-ahmed/2/orig -> origin/gh/syed-ahmed/2/orig 2025-07-17T08:34:51.4016528Z * [new branch] gh/teja-rao/1/base -> origin/gh/teja-rao/1/base 2025-07-17T08:34:51.4019923Z * [new branch] gh/teja-rao/1/head -> origin/gh/teja-rao/1/head 2025-07-17T08:34:51.4023331Z * [new branch] gh/teja-rao/1/orig -> origin/gh/teja-rao/1/orig 2025-07-17T08:34:51.4028107Z * [new branch] gh/teja-rao/2/base -> origin/gh/teja-rao/2/base 2025-07-17T08:34:51.4031348Z * [new branch] gh/teja-rao/2/head -> origin/gh/teja-rao/2/head 2025-07-17T08:34:51.4034851Z * [new branch] gh/teja-rao/2/orig -> origin/gh/teja-rao/2/orig 2025-07-17T08:34:51.4040456Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-07-17T08:34:51.4043778Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-07-17T08:34:51.4047107Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-07-17T08:34:51.4051703Z * [new branch] gh/tianyu-l/3/base -> origin/gh/tianyu-l/3/base 2025-07-17T08:34:51.4055171Z * [new branch] gh/tianyu-l/3/orig -> origin/gh/tianyu-l/3/orig 2025-07-17T08:34:51.4059715Z * [new branch] gh/tianyu-l/4/base -> origin/gh/tianyu-l/4/base 2025-07-17T08:34:51.4063027Z * [new branch] gh/tianyu-l/4/head -> origin/gh/tianyu-l/4/head 2025-07-17T08:34:51.4066403Z * [new branch] gh/tianyu-l/4/orig -> origin/gh/tianyu-l/4/orig 2025-07-17T08:34:51.4071929Z * [new branch] gh/titaiwangms/1/base -> origin/gh/titaiwangms/1/base 2025-07-17T08:34:51.4075450Z * [new branch] gh/titaiwangms/1/head -> origin/gh/titaiwangms/1/head 2025-07-17T08:34:51.4078813Z * [new branch] gh/titaiwangms/1/orig -> origin/gh/titaiwangms/1/orig 2025-07-17T08:34:51.4083330Z * [new branch] gh/titaiwangms/2/base -> origin/gh/titaiwangms/2/base 2025-07-17T08:34:51.4086739Z * [new branch] gh/titaiwangms/2/head -> origin/gh/titaiwangms/2/head 2025-07-17T08:34:51.4090091Z * [new branch] gh/titaiwangms/2/orig -> origin/gh/titaiwangms/2/orig 2025-07-17T08:34:51.4094623Z * [new branch] gh/titaiwangms/3/base -> origin/gh/titaiwangms/3/base 2025-07-17T08:34:51.4098095Z * [new branch] gh/titaiwangms/3/head -> origin/gh/titaiwangms/3/head 2025-07-17T08:34:51.4101649Z * [new branch] gh/titaiwangms/3/orig -> origin/gh/titaiwangms/3/orig 2025-07-17T08:34:51.4106315Z * [new branch] gh/titaiwangms/4/base -> origin/gh/titaiwangms/4/base 2025-07-17T08:34:51.4109691Z * [new branch] gh/titaiwangms/4/head -> origin/gh/titaiwangms/4/head 2025-07-17T08:34:51.4113052Z * [new branch] gh/titaiwangms/4/orig -> origin/gh/titaiwangms/4/orig 2025-07-17T08:34:51.4117750Z * [new branch] gh/titaiwangms/5/base -> origin/gh/titaiwangms/5/base 2025-07-17T08:34:51.4121054Z * [new branch] gh/titaiwangms/5/head -> origin/gh/titaiwangms/5/head 2025-07-17T08:34:51.4124462Z * [new branch] gh/titaiwangms/5/orig -> origin/gh/titaiwangms/5/orig 2025-07-17T08:34:51.4129043Z * [new branch] gh/titaiwangms/6/base -> origin/gh/titaiwangms/6/base 2025-07-17T08:34:51.4132413Z * [new branch] gh/titaiwangms/6/head -> origin/gh/titaiwangms/6/head 2025-07-17T08:34:51.4135781Z * [new branch] gh/titaiwangms/6/orig -> origin/gh/titaiwangms/6/orig 2025-07-17T08:34:51.4140338Z * [new branch] gh/titaiwangms/7/base -> origin/gh/titaiwangms/7/base 2025-07-17T08:34:51.4143786Z * [new branch] gh/titaiwangms/7/head -> origin/gh/titaiwangms/7/head 2025-07-17T08:34:51.4147134Z * [new branch] gh/titaiwangms/7/orig -> origin/gh/titaiwangms/7/orig 2025-07-17T08:34:51.4151842Z * [new branch] gh/titaiwangms/8/base -> origin/gh/titaiwangms/8/base 2025-07-17T08:34:51.4155351Z * [new branch] gh/titaiwangms/8/head -> origin/gh/titaiwangms/8/head 2025-07-17T08:34:51.4158948Z * [new branch] gh/titaiwangms/8/orig -> origin/gh/titaiwangms/8/orig 2025-07-17T08:34:51.4164368Z * [new branch] gh/tugsbayasgalan/301/base -> origin/gh/tugsbayasgalan/301/base 2025-07-17T08:34:51.4167776Z * [new branch] gh/tugsbayasgalan/301/head -> origin/gh/tugsbayasgalan/301/head 2025-07-17T08:34:51.4171180Z * [new branch] gh/tugsbayasgalan/301/orig -> origin/gh/tugsbayasgalan/301/orig 2025-07-17T08:34:51.4176704Z * [new branch] gh/v0i0/1/base -> origin/gh/v0i0/1/base 2025-07-17T08:34:51.4180056Z * [new branch] gh/v0i0/1/head -> origin/gh/v0i0/1/head 2025-07-17T08:34:51.4183417Z * [new branch] gh/v0i0/1/orig -> origin/gh/v0i0/1/orig 2025-07-17T08:34:51.4188975Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-07-17T08:34:51.4193505Z * [new branch] gh/vkuzo/10/base -> origin/gh/vkuzo/10/base 2025-07-17T08:34:51.4197077Z * [new branch] gh/vkuzo/10/head -> origin/gh/vkuzo/10/head 2025-07-17T08:34:51.4200433Z * [new branch] gh/vkuzo/10/orig -> origin/gh/vkuzo/10/orig 2025-07-17T08:34:51.4205021Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-07-17T08:34:51.4209610Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-07-17T08:34:51.4215101Z * [new branch] gh/wanchaol/373/base -> origin/gh/wanchaol/373/base 2025-07-17T08:34:51.4218431Z * [new branch] gh/wanchaol/373/head -> origin/gh/wanchaol/373/head 2025-07-17T08:34:51.4221836Z * [new branch] gh/wanchaol/373/orig -> origin/gh/wanchaol/373/orig 2025-07-17T08:34:51.4226481Z * [new branch] gh/wanchaol/374/base -> origin/gh/wanchaol/374/base 2025-07-17T08:34:51.4229886Z * [new branch] gh/wanchaol/374/head -> origin/gh/wanchaol/374/head 2025-07-17T08:34:51.4233351Z * [new branch] gh/wanchaol/374/orig -> origin/gh/wanchaol/374/orig 2025-07-17T08:34:51.4239409Z * [new branch] gh/wconstab/392/base -> origin/gh/wconstab/392/base 2025-07-17T08:34:51.4242804Z * [new branch] gh/wconstab/392/head -> origin/gh/wconstab/392/head 2025-07-17T08:34:51.4246202Z * [new branch] gh/wconstab/392/orig -> origin/gh/wconstab/392/orig 2025-07-17T08:34:51.4250936Z * [new branch] gh/wconstab/419/base -> origin/gh/wconstab/419/base 2025-07-17T08:34:51.4255205Z * [new branch] gh/wconstab/419/head -> origin/gh/wconstab/419/head 2025-07-17T08:34:51.4257660Z * [new branch] gh/wconstab/419/orig -> origin/gh/wconstab/419/orig 2025-07-17T08:34:51.4262258Z * [new branch] gh/wconstab/420/base -> origin/gh/wconstab/420/base 2025-07-17T08:34:51.4265671Z * [new branch] gh/wconstab/420/head -> origin/gh/wconstab/420/head 2025-07-17T08:34:51.4269202Z * [new branch] gh/wconstab/420/orig -> origin/gh/wconstab/420/orig 2025-07-17T08:34:51.4273721Z * [new branch] gh/wconstab/421/base -> origin/gh/wconstab/421/base 2025-07-17T08:34:51.4277502Z * [new branch] gh/wconstab/421/orig -> origin/gh/wconstab/421/orig 2025-07-17T08:34:51.4282050Z * [new branch] gh/wconstab/422/base -> origin/gh/wconstab/422/base 2025-07-17T08:34:51.4285510Z * [new branch] gh/wconstab/422/head -> origin/gh/wconstab/422/head 2025-07-17T08:34:51.4289011Z * [new branch] gh/wconstab/422/orig -> origin/gh/wconstab/422/orig 2025-07-17T08:34:51.4293706Z * [new branch] gh/wconstab/423/base -> origin/gh/wconstab/423/base 2025-07-17T08:34:51.4297063Z * [new branch] gh/wconstab/423/head -> origin/gh/wconstab/423/head 2025-07-17T08:34:51.4300682Z * [new branch] gh/wconstab/423/orig -> origin/gh/wconstab/423/orig 2025-07-17T08:34:51.4305524Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-07-17T08:34:51.4308635Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-07-17T08:34:51.4311935Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-07-17T08:34:51.4318069Z * [new branch] gh/wconstab/425/base -> origin/gh/wconstab/425/base 2025-07-17T08:34:51.4321465Z * [new branch] gh/wconstab/425/head -> origin/gh/wconstab/425/head 2025-07-17T08:34:51.4324872Z * [new branch] gh/wconstab/425/orig -> origin/gh/wconstab/425/orig 2025-07-17T08:34:51.4329512Z * [new branch] gh/wconstab/426/base -> origin/gh/wconstab/426/base 2025-07-17T08:34:51.4332858Z * [new branch] gh/wconstab/426/head -> origin/gh/wconstab/426/head 2025-07-17T08:34:51.4336215Z * [new branch] gh/wconstab/426/orig -> origin/gh/wconstab/426/orig 2025-07-17T08:34:51.4340608Z * [new branch] gh/wconstab/427/base -> origin/gh/wconstab/427/base 2025-07-17T08:34:51.4343917Z * [new branch] gh/wconstab/427/head -> origin/gh/wconstab/427/head 2025-07-17T08:34:51.4347327Z * [new branch] gh/wconstab/427/orig -> origin/gh/wconstab/427/orig 2025-07-17T08:34:51.4351718Z * [new branch] gh/wconstab/428/base -> origin/gh/wconstab/428/base 2025-07-17T08:34:51.4355212Z * [new branch] gh/wconstab/428/head -> origin/gh/wconstab/428/head 2025-07-17T08:34:51.4358623Z * [new branch] gh/wconstab/428/orig -> origin/gh/wconstab/428/orig 2025-07-17T08:34:51.4363089Z * [new branch] gh/wconstab/429/base -> origin/gh/wconstab/429/base 2025-07-17T08:34:51.4366483Z * [new branch] gh/wconstab/429/head -> origin/gh/wconstab/429/head 2025-07-17T08:34:51.4369869Z * [new branch] gh/wconstab/429/orig -> origin/gh/wconstab/429/orig 2025-07-17T08:34:51.4374428Z * [new branch] gh/wconstab/430/base -> origin/gh/wconstab/430/base 2025-07-17T08:34:51.4377712Z * [new branch] gh/wconstab/430/head -> origin/gh/wconstab/430/head 2025-07-17T08:34:51.4381097Z * [new branch] gh/wconstab/430/orig -> origin/gh/wconstab/430/orig 2025-07-17T08:34:51.4385459Z * [new branch] gh/wconstab/431/base -> origin/gh/wconstab/431/base 2025-07-17T08:34:51.4388758Z * [new branch] gh/wconstab/431/head -> origin/gh/wconstab/431/head 2025-07-17T08:34:51.4392095Z * [new branch] gh/wconstab/431/orig -> origin/gh/wconstab/431/orig 2025-07-17T08:34:51.4397839Z * [new branch] gh/weifengpy/27/base -> origin/gh/weifengpy/27/base 2025-07-17T08:34:51.4401068Z * [new branch] gh/weifengpy/27/head -> origin/gh/weifengpy/27/head 2025-07-17T08:34:51.4404511Z * [new branch] gh/weifengpy/27/orig -> origin/gh/weifengpy/27/orig 2025-07-17T08:34:51.4409014Z * [new branch] gh/weifengpy/29/base -> origin/gh/weifengpy/29/base 2025-07-17T08:34:51.4412336Z * [new branch] gh/weifengpy/29/head -> origin/gh/weifengpy/29/head 2025-07-17T08:34:51.4415615Z * [new branch] gh/weifengpy/29/orig -> origin/gh/weifengpy/29/orig 2025-07-17T08:34:51.4420162Z * [new branch] gh/weifengpy/30/base -> origin/gh/weifengpy/30/base 2025-07-17T08:34:51.4423473Z * [new branch] gh/weifengpy/30/head -> origin/gh/weifengpy/30/head 2025-07-17T08:34:51.4426788Z * [new branch] gh/weifengpy/30/orig -> origin/gh/weifengpy/30/orig 2025-07-17T08:34:51.4432322Z * [new branch] gh/williamwen42/196/base -> origin/gh/williamwen42/196/base 2025-07-17T08:34:51.4435852Z * [new branch] gh/williamwen42/196/head -> origin/gh/williamwen42/196/head 2025-07-17T08:34:51.4439391Z * [new branch] gh/williamwen42/196/orig -> origin/gh/williamwen42/196/orig 2025-07-17T08:34:51.4443795Z * [new branch] gh/williamwen42/209/base -> origin/gh/williamwen42/209/base 2025-07-17T08:34:51.4447234Z * [new branch] gh/williamwen42/209/head -> origin/gh/williamwen42/209/head 2025-07-17T08:34:51.4450581Z * [new branch] gh/williamwen42/209/orig -> origin/gh/williamwen42/209/orig 2025-07-17T08:34:51.4455124Z * [new branch] gh/williamwen42/241/base -> origin/gh/williamwen42/241/base 2025-07-17T08:34:51.4458524Z * [new branch] gh/williamwen42/241/head -> origin/gh/williamwen42/241/head 2025-07-17T08:34:51.4461886Z * [new branch] gh/williamwen42/241/orig -> origin/gh/williamwen42/241/orig 2025-07-17T08:34:51.4466522Z * [new branch] gh/williamwen42/242/base -> origin/gh/williamwen42/242/base 2025-07-17T08:34:51.4470021Z * [new branch] gh/williamwen42/242/head -> origin/gh/williamwen42/242/head 2025-07-17T08:34:51.4473407Z * [new branch] gh/williamwen42/242/orig -> origin/gh/williamwen42/242/orig 2025-07-17T08:34:51.4478634Z * [new branch] gh/williamwen42/243/base -> origin/gh/williamwen42/243/base 2025-07-17T08:34:51.4482064Z * [new branch] gh/williamwen42/243/head -> origin/gh/williamwen42/243/head 2025-07-17T08:34:51.4485443Z * [new branch] gh/williamwen42/243/orig -> origin/gh/williamwen42/243/orig 2025-07-17T08:34:51.4490167Z * [new branch] gh/williamwen42/244/base -> origin/gh/williamwen42/244/base 2025-07-17T08:34:51.4493951Z * [new branch] gh/williamwen42/244/head -> origin/gh/williamwen42/244/head 2025-07-17T08:34:51.4497308Z * [new branch] gh/williamwen42/244/orig -> origin/gh/williamwen42/244/orig 2025-07-17T08:34:51.4501946Z * [new branch] gh/williamwen42/245/base -> origin/gh/williamwen42/245/base 2025-07-17T08:34:51.4505363Z * [new branch] gh/williamwen42/245/head -> origin/gh/williamwen42/245/head 2025-07-17T08:34:51.4508775Z * [new branch] gh/williamwen42/245/orig -> origin/gh/williamwen42/245/orig 2025-07-17T08:34:51.4513569Z * [new branch] gh/williamwen42/246/base -> origin/gh/williamwen42/246/base 2025-07-17T08:34:51.4517126Z * [new branch] gh/williamwen42/246/head -> origin/gh/williamwen42/246/head 2025-07-17T08:34:51.4520462Z * [new branch] gh/williamwen42/246/orig -> origin/gh/williamwen42/246/orig 2025-07-17T08:34:51.4525139Z * [new branch] gh/williamwen42/248/base -> origin/gh/williamwen42/248/base 2025-07-17T08:34:51.4528524Z * [new branch] gh/williamwen42/248/head -> origin/gh/williamwen42/248/head 2025-07-17T08:34:51.4531975Z * [new branch] gh/williamwen42/248/orig -> origin/gh/williamwen42/248/orig 2025-07-17T08:34:51.4536584Z * [new branch] gh/williamwen42/249/base -> origin/gh/williamwen42/249/base 2025-07-17T08:34:51.4540097Z * [new branch] gh/williamwen42/249/head -> origin/gh/williamwen42/249/head 2025-07-17T08:34:51.4543446Z * [new branch] gh/williamwen42/249/orig -> origin/gh/williamwen42/249/orig 2025-07-17T08:34:51.4547840Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-07-17T08:34:51.4551206Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-07-17T08:34:51.4557670Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-07-17T08:34:51.4562944Z * [new branch] gh/williamwen42/251/base -> origin/gh/williamwen42/251/base 2025-07-17T08:34:51.4566416Z * [new branch] gh/williamwen42/251/head -> origin/gh/williamwen42/251/head 2025-07-17T08:34:51.4569772Z * [new branch] gh/williamwen42/251/orig -> origin/gh/williamwen42/251/orig 2025-07-17T08:34:51.4574707Z * [new branch] gh/williamwen42/252/base -> origin/gh/williamwen42/252/base 2025-07-17T08:34:51.4577935Z * [new branch] gh/williamwen42/252/head -> origin/gh/williamwen42/252/head 2025-07-17T08:34:51.4581297Z * [new branch] gh/williamwen42/252/orig -> origin/gh/williamwen42/252/orig 2025-07-17T08:34:51.4585840Z * [new branch] gh/williamwen42/253/base -> origin/gh/williamwen42/253/base 2025-07-17T08:34:51.4589310Z * [new branch] gh/williamwen42/253/head -> origin/gh/williamwen42/253/head 2025-07-17T08:34:51.4592836Z * [new branch] gh/williamwen42/253/orig -> origin/gh/williamwen42/253/orig 2025-07-17T08:34:51.4597673Z * [new branch] gh/williamwen42/254/base -> origin/gh/williamwen42/254/base 2025-07-17T08:34:51.4600978Z * [new branch] gh/williamwen42/254/head -> origin/gh/williamwen42/254/head 2025-07-17T08:34:51.4604345Z * [new branch] gh/williamwen42/254/orig -> origin/gh/williamwen42/254/orig 2025-07-17T08:34:51.4609137Z * [new branch] gh/williamwen42/255/base -> origin/gh/williamwen42/255/base 2025-07-17T08:34:51.4612483Z * [new branch] gh/williamwen42/255/head -> origin/gh/williamwen42/255/head 2025-07-17T08:34:51.4615794Z * [new branch] gh/williamwen42/255/orig -> origin/gh/williamwen42/255/orig 2025-07-17T08:34:51.4620205Z * [new branch] gh/williamwen42/256/base -> origin/gh/williamwen42/256/base 2025-07-17T08:34:51.4623569Z * [new branch] gh/williamwen42/256/head -> origin/gh/williamwen42/256/head 2025-07-17T08:34:51.4626887Z * [new branch] gh/williamwen42/256/orig -> origin/gh/williamwen42/256/orig 2025-07-17T08:34:51.4631520Z * [new branch] gh/williamwen42/257/base -> origin/gh/williamwen42/257/base 2025-07-17T08:34:51.4635039Z * [new branch] gh/williamwen42/257/head -> origin/gh/williamwen42/257/head 2025-07-17T08:34:51.4638524Z * [new branch] gh/williamwen42/257/orig -> origin/gh/williamwen42/257/orig 2025-07-17T08:34:51.4642914Z * [new branch] gh/williamwen42/258/base -> origin/gh/williamwen42/258/base 2025-07-17T08:34:51.4646279Z * [new branch] gh/williamwen42/258/head -> origin/gh/williamwen42/258/head 2025-07-17T08:34:51.4649688Z * [new branch] gh/williamwen42/258/orig -> origin/gh/williamwen42/258/orig 2025-07-17T08:34:51.4654227Z * [new branch] gh/williamwen42/259/base -> origin/gh/williamwen42/259/base 2025-07-17T08:34:51.4657596Z * [new branch] gh/williamwen42/259/head -> origin/gh/williamwen42/259/head 2025-07-17T08:34:51.4661026Z * [new branch] gh/williamwen42/259/orig -> origin/gh/williamwen42/259/orig 2025-07-17T08:34:51.4665491Z * [new branch] gh/williamwen42/260/base -> origin/gh/williamwen42/260/base 2025-07-17T08:34:51.4668836Z * [new branch] gh/williamwen42/260/head -> origin/gh/williamwen42/260/head 2025-07-17T08:34:51.4672151Z * [new branch] gh/williamwen42/260/orig -> origin/gh/williamwen42/260/orig 2025-07-17T08:34:51.4676967Z * [new branch] gh/williamwen42/261/base -> origin/gh/williamwen42/261/base 2025-07-17T08:34:51.4680258Z * [new branch] gh/williamwen42/261/head -> origin/gh/williamwen42/261/head 2025-07-17T08:34:51.4683681Z * [new branch] gh/williamwen42/261/orig -> origin/gh/williamwen42/261/orig 2025-07-17T08:34:51.4688372Z * [new branch] gh/williamwen42/262/base -> origin/gh/williamwen42/262/base 2025-07-17T08:34:51.4691729Z * [new branch] gh/williamwen42/262/head -> origin/gh/williamwen42/262/head 2025-07-17T08:34:51.4695100Z * [new branch] gh/williamwen42/262/orig -> origin/gh/williamwen42/262/orig 2025-07-17T08:34:51.4699990Z * [new branch] gh/williamwen42/263/base -> origin/gh/williamwen42/263/base 2025-07-17T08:34:51.4703164Z * [new branch] gh/williamwen42/263/head -> origin/gh/williamwen42/263/head 2025-07-17T08:34:51.4706457Z * [new branch] gh/williamwen42/263/orig -> origin/gh/williamwen42/263/orig 2025-07-17T08:34:51.4711938Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-07-17T08:34:51.4715575Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-07-17T08:34:51.4720033Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-07-17T08:34:51.4723260Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-07-17T08:34:51.4728007Z * [new branch] gh/xmfan/18/base -> origin/gh/xmfan/18/base 2025-07-17T08:34:51.4731323Z * [new branch] gh/xmfan/18/head -> origin/gh/xmfan/18/head 2025-07-17T08:34:51.4735919Z * [new branch] gh/xmfan/194/base -> origin/gh/xmfan/194/base 2025-07-17T08:34:51.4739258Z * [new branch] gh/xmfan/194/head -> origin/gh/xmfan/194/head 2025-07-17T08:34:51.4742554Z * [new branch] gh/xmfan/194/orig -> origin/gh/xmfan/194/orig 2025-07-17T08:34:51.4747124Z * [new branch] gh/xmfan/206/base -> origin/gh/xmfan/206/base 2025-07-17T08:34:51.4750366Z * [new branch] gh/xmfan/206/head -> origin/gh/xmfan/206/head 2025-07-17T08:34:51.4753736Z * [new branch] gh/xmfan/206/orig -> origin/gh/xmfan/206/orig 2025-07-17T08:34:51.4758499Z * [new branch] gh/xmfan/213/base -> origin/gh/xmfan/213/base 2025-07-17T08:34:51.4761783Z * [new branch] gh/xmfan/213/head -> origin/gh/xmfan/213/head 2025-07-17T08:34:51.4765121Z * [new branch] gh/xmfan/213/orig -> origin/gh/xmfan/213/orig 2025-07-17T08:34:51.4769758Z * [new branch] gh/xmfan/223/base -> origin/gh/xmfan/223/base 2025-07-17T08:34:51.4773065Z * [new branch] gh/xmfan/223/head -> origin/gh/xmfan/223/head 2025-07-17T08:34:51.4776487Z * [new branch] gh/xmfan/223/orig -> origin/gh/xmfan/223/orig 2025-07-17T08:34:51.4781059Z * [new branch] gh/xmfan/225/base -> origin/gh/xmfan/225/base 2025-07-17T08:34:51.4784396Z * [new branch] gh/xmfan/225/head -> origin/gh/xmfan/225/head 2025-07-17T08:34:51.4787703Z * [new branch] gh/xmfan/225/orig -> origin/gh/xmfan/225/orig 2025-07-17T08:34:51.4792268Z * [new branch] gh/xmfan/228/base -> origin/gh/xmfan/228/base 2025-07-17T08:34:51.4795749Z * [new branch] gh/xmfan/228/head -> origin/gh/xmfan/228/head 2025-07-17T08:34:51.4799065Z * [new branch] gh/xmfan/228/orig -> origin/gh/xmfan/228/orig 2025-07-17T08:34:51.4803527Z * [new branch] gh/xmfan/229/base -> origin/gh/xmfan/229/base 2025-07-17T08:34:51.4806927Z * [new branch] gh/xmfan/229/head -> origin/gh/xmfan/229/head 2025-07-17T08:34:51.4810218Z * [new branch] gh/xmfan/229/orig -> origin/gh/xmfan/229/orig 2025-07-17T08:34:51.4814883Z * [new branch] gh/xmfan/230/base -> origin/gh/xmfan/230/base 2025-07-17T08:34:51.4818211Z * [new branch] gh/xmfan/230/head -> origin/gh/xmfan/230/head 2025-07-17T08:34:51.4821530Z * [new branch] gh/xmfan/230/orig -> origin/gh/xmfan/230/orig 2025-07-17T08:34:51.4826033Z * [new branch] gh/xmfan/237/base -> origin/gh/xmfan/237/base 2025-07-17T08:34:51.4829375Z * [new branch] gh/xmfan/237/head -> origin/gh/xmfan/237/head 2025-07-17T08:34:51.4832793Z * [new branch] gh/xmfan/237/orig -> origin/gh/xmfan/237/orig 2025-07-17T08:34:51.4837811Z * [new branch] gh/xmfan/243/base -> origin/gh/xmfan/243/base 2025-07-17T08:34:51.4840900Z * [new branch] gh/xmfan/243/head -> origin/gh/xmfan/243/head 2025-07-17T08:34:51.4844247Z * [new branch] gh/xmfan/243/orig -> origin/gh/xmfan/243/orig 2025-07-17T08:34:51.4848793Z * [new branch] gh/xmfan/244/base -> origin/gh/xmfan/244/base 2025-07-17T08:34:51.4852115Z * [new branch] gh/xmfan/244/head -> origin/gh/xmfan/244/head 2025-07-17T08:34:51.4855479Z * [new branch] gh/xmfan/244/orig -> origin/gh/xmfan/244/orig 2025-07-17T08:34:51.4860205Z * [new branch] gh/xmfan/246/base -> origin/gh/xmfan/246/base 2025-07-17T08:34:51.4863529Z * [new branch] gh/xmfan/246/head -> origin/gh/xmfan/246/head 2025-07-17T08:34:51.4866848Z * [new branch] gh/xmfan/246/orig -> origin/gh/xmfan/246/orig 2025-07-17T08:34:51.4871505Z * [new branch] gh/xmfan/248/base -> origin/gh/xmfan/248/base 2025-07-17T08:34:51.4874948Z * [new branch] gh/xmfan/248/head -> origin/gh/xmfan/248/head 2025-07-17T08:34:51.4878654Z * [new branch] gh/xmfan/248/orig -> origin/gh/xmfan/248/orig 2025-07-17T08:34:51.4883278Z * [new branch] gh/xmfan/250/base -> origin/gh/xmfan/250/base 2025-07-17T08:34:51.4886599Z * [new branch] gh/xmfan/250/head -> origin/gh/xmfan/250/head 2025-07-17T08:34:51.4889955Z * [new branch] gh/xmfan/250/orig -> origin/gh/xmfan/250/orig 2025-07-17T08:34:51.4894511Z * [new branch] gh/xmfan/253/base -> origin/gh/xmfan/253/base 2025-07-17T08:34:51.4897841Z * [new branch] gh/xmfan/253/head -> origin/gh/xmfan/253/head 2025-07-17T08:34:51.4901198Z * [new branch] gh/xmfan/253/orig -> origin/gh/xmfan/253/orig 2025-07-17T08:34:51.4905770Z * [new branch] gh/xmfan/254/base -> origin/gh/xmfan/254/base 2025-07-17T08:34:51.4909080Z * [new branch] gh/xmfan/254/head -> origin/gh/xmfan/254/head 2025-07-17T08:34:51.4912419Z * [new branch] gh/xmfan/254/orig -> origin/gh/xmfan/254/orig 2025-07-17T08:34:51.4917188Z * [new branch] gh/xmfan/255/base -> origin/gh/xmfan/255/base 2025-07-17T08:34:51.4920448Z * [new branch] gh/xmfan/255/head -> origin/gh/xmfan/255/head 2025-07-17T08:34:51.4923801Z * [new branch] gh/xmfan/255/orig -> origin/gh/xmfan/255/orig 2025-07-17T08:34:51.4928326Z * [new branch] gh/xmfan/256/base -> origin/gh/xmfan/256/base 2025-07-17T08:34:51.4931677Z * [new branch] gh/xmfan/256/head -> origin/gh/xmfan/256/head 2025-07-17T08:34:51.4935062Z * [new branch] gh/xmfan/256/orig -> origin/gh/xmfan/256/orig 2025-07-17T08:34:51.4939525Z * [new branch] gh/xmfan/257/base -> origin/gh/xmfan/257/base 2025-07-17T08:34:51.4942831Z * [new branch] gh/xmfan/257/head -> origin/gh/xmfan/257/head 2025-07-17T08:34:51.4946178Z * [new branch] gh/xmfan/257/orig -> origin/gh/xmfan/257/orig 2025-07-17T08:34:51.4950790Z * [new branch] gh/xmfan/258/base -> origin/gh/xmfan/258/base 2025-07-17T08:34:51.4954204Z * [new branch] gh/xmfan/258/head -> origin/gh/xmfan/258/head 2025-07-17T08:34:51.4957733Z * [new branch] gh/xmfan/258/orig -> origin/gh/xmfan/258/orig 2025-07-17T08:34:51.4962271Z * [new branch] gh/xmfan/259/base -> origin/gh/xmfan/259/base 2025-07-17T08:34:51.4965613Z * [new branch] gh/xmfan/259/head -> origin/gh/xmfan/259/head 2025-07-17T08:34:51.4968932Z * [new branch] gh/xmfan/259/orig -> origin/gh/xmfan/259/orig 2025-07-17T08:34:51.4973680Z * [new branch] gh/xmfan/260/base -> origin/gh/xmfan/260/base 2025-07-17T08:34:51.4976843Z * [new branch] gh/xmfan/260/head -> origin/gh/xmfan/260/head 2025-07-17T08:34:51.4980181Z * [new branch] gh/xmfan/260/orig -> origin/gh/xmfan/260/orig 2025-07-17T08:34:51.4984763Z * [new branch] gh/xmfan/261/base -> origin/gh/xmfan/261/base 2025-07-17T08:34:51.4988131Z * [new branch] gh/xmfan/261/head -> origin/gh/xmfan/261/head 2025-07-17T08:34:51.4991436Z * [new branch] gh/xmfan/261/orig -> origin/gh/xmfan/261/orig 2025-07-17T08:34:51.4996383Z * [new branch] gh/xmfan/262/base -> origin/gh/xmfan/262/base 2025-07-17T08:34:51.4999707Z * [new branch] gh/xmfan/262/head -> origin/gh/xmfan/262/head 2025-07-17T08:34:51.5002990Z * [new branch] gh/xmfan/262/orig -> origin/gh/xmfan/262/orig 2025-07-17T08:34:51.5007527Z * [new branch] gh/xmfan/263/base -> origin/gh/xmfan/263/base 2025-07-17T08:34:51.5010822Z * [new branch] gh/xmfan/263/head -> origin/gh/xmfan/263/head 2025-07-17T08:34:51.5014204Z * [new branch] gh/xmfan/263/orig -> origin/gh/xmfan/263/orig 2025-07-17T08:34:51.5018739Z * [new branch] gh/xmfan/264/base -> origin/gh/xmfan/264/base 2025-07-17T08:34:51.5022014Z * [new branch] gh/xmfan/264/head -> origin/gh/xmfan/264/head 2025-07-17T08:34:51.5025312Z * [new branch] gh/xmfan/264/orig -> origin/gh/xmfan/264/orig 2025-07-17T08:34:51.5029861Z * [new branch] gh/xmfan/265/base -> origin/gh/xmfan/265/base 2025-07-17T08:34:51.5033104Z * [new branch] gh/xmfan/265/head -> origin/gh/xmfan/265/head 2025-07-17T08:34:51.5036629Z * [new branch] gh/xmfan/265/orig -> origin/gh/xmfan/265/orig 2025-07-17T08:34:51.5041216Z * [new branch] gh/xmfan/266/base -> origin/gh/xmfan/266/base 2025-07-17T08:34:51.5044652Z * [new branch] gh/xmfan/266/head -> origin/gh/xmfan/266/head 2025-07-17T08:34:51.5047944Z * [new branch] gh/xmfan/266/orig -> origin/gh/xmfan/266/orig 2025-07-17T08:34:51.5052661Z * [new branch] gh/xmfan/267/base -> origin/gh/xmfan/267/base 2025-07-17T08:34:51.5055956Z * [new branch] gh/xmfan/267/head -> origin/gh/xmfan/267/head 2025-07-17T08:34:51.5059274Z * [new branch] gh/xmfan/267/orig -> origin/gh/xmfan/267/orig 2025-07-17T08:34:51.5063790Z * [new branch] gh/xmfan/268/base -> origin/gh/xmfan/268/base 2025-07-17T08:34:51.5067377Z * [new branch] gh/xmfan/268/head -> origin/gh/xmfan/268/head 2025-07-17T08:34:51.5070731Z * [new branch] gh/xmfan/268/orig -> origin/gh/xmfan/268/orig 2025-07-17T08:34:51.5075444Z * [new branch] gh/xmfan/269/base -> origin/gh/xmfan/269/base 2025-07-17T08:34:51.5078746Z * [new branch] gh/xmfan/269/head -> origin/gh/xmfan/269/head 2025-07-17T08:34:51.5082086Z * [new branch] gh/xmfan/269/orig -> origin/gh/xmfan/269/orig 2025-07-17T08:34:51.5087719Z * [new branch] gh/xuanzhang816/12/base -> origin/gh/xuanzhang816/12/base 2025-07-17T08:34:51.5091517Z * [new branch] gh/xuanzhang816/12/head -> origin/gh/xuanzhang816/12/head 2025-07-17T08:34:51.5094647Z * [new branch] gh/xuanzhang816/12/orig -> origin/gh/xuanzhang816/12/orig 2025-07-17T08:34:51.5099149Z * [new branch] gh/xuanzhang816/14/base -> origin/gh/xuanzhang816/14/base 2025-07-17T08:34:51.5102486Z * [new branch] gh/xuanzhang816/14/head -> origin/gh/xuanzhang816/14/head 2025-07-17T08:34:51.5105797Z * [new branch] gh/xuanzhang816/14/orig -> origin/gh/xuanzhang816/14/orig 2025-07-17T08:34:51.5110884Z * [new branch] gh/xuanzhang816/15/base -> origin/gh/xuanzhang816/15/base 2025-07-17T08:34:51.5113727Z * [new branch] gh/xuanzhang816/15/head -> origin/gh/xuanzhang816/15/head 2025-07-17T08:34:51.5117316Z * [new branch] gh/xuanzhang816/15/orig -> origin/gh/xuanzhang816/15/orig 2025-07-17T08:34:51.5121764Z * [new branch] gh/xuanzhang816/16/base -> origin/gh/xuanzhang816/16/base 2025-07-17T08:34:51.5125109Z * [new branch] gh/xuanzhang816/16/head -> origin/gh/xuanzhang816/16/head 2025-07-17T08:34:51.5128484Z * [new branch] gh/xuanzhang816/16/orig -> origin/gh/xuanzhang816/16/orig 2025-07-17T08:34:51.5133135Z * [new branch] gh/xuanzhang816/17/base -> origin/gh/xuanzhang816/17/base 2025-07-17T08:34:51.5136498Z * [new branch] gh/xuanzhang816/17/head -> origin/gh/xuanzhang816/17/head 2025-07-17T08:34:51.5139862Z * [new branch] gh/xuanzhang816/17/orig -> origin/gh/xuanzhang816/17/orig 2025-07-17T08:34:51.5144385Z * [new branch] gh/xuanzhang816/18/base -> origin/gh/xuanzhang816/18/base 2025-07-17T08:34:51.5147855Z * [new branch] gh/xuanzhang816/18/head -> origin/gh/xuanzhang816/18/head 2025-07-17T08:34:51.5151179Z * [new branch] gh/xuanzhang816/18/orig -> origin/gh/xuanzhang816/18/orig 2025-07-17T08:34:51.5156094Z * [new branch] gh/xuanzhang816/19/base -> origin/gh/xuanzhang816/19/base 2025-07-17T08:34:51.5159294Z * [new branch] gh/xuanzhang816/19/head -> origin/gh/xuanzhang816/19/head 2025-07-17T08:34:51.5162666Z * [new branch] gh/xuanzhang816/19/orig -> origin/gh/xuanzhang816/19/orig 2025-07-17T08:34:51.5167390Z * [new branch] gh/xuanzhang816/20/base -> origin/gh/xuanzhang816/20/base 2025-07-17T08:34:51.5170771Z * [new branch] gh/xuanzhang816/20/head -> origin/gh/xuanzhang816/20/head 2025-07-17T08:34:51.5174139Z * [new branch] gh/xuanzhang816/20/orig -> origin/gh/xuanzhang816/20/orig 2025-07-17T08:34:51.5179735Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-07-17T08:34:51.5183052Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-07-17T08:34:51.5186499Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-07-17T08:34:51.5191070Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-07-17T08:34:51.5194545Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-07-17T08:34:51.5198015Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-07-17T08:34:51.5202518Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-07-17T08:34:51.5205853Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-07-17T08:34:51.5209278Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-07-17T08:34:51.5213784Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-07-17T08:34:51.5217234Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-07-17T08:34:51.5220726Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-07-17T08:34:51.5225172Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-07-17T08:34:51.5228563Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-07-17T08:34:51.5231909Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-07-17T08:34:51.5236590Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-07-17T08:34:51.5240119Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-07-17T08:34:51.5243222Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-07-17T08:34:51.5247825Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-07-17T08:34:51.5251173Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-07-17T08:34:51.5254523Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-07-17T08:34:51.5259399Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-07-17T08:34:51.5262774Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-07-17T08:34:51.5266172Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-07-17T08:34:51.5270700Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-07-17T08:34:51.5274107Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-07-17T08:34:51.5278815Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-07-17T08:34:51.5282124Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-07-17T08:34:51.5285411Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-07-17T08:34:51.5289979Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-07-17T08:34:51.5293327Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-07-17T08:34:51.5296638Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-07-17T08:34:51.5301149Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-07-17T08:34:51.5304564Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-07-17T08:34:51.5307842Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-07-17T08:34:51.5312548Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-07-17T08:34:51.5316161Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-07-17T08:34:51.5319425Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-07-17T08:34:51.5323977Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-07-17T08:34:51.5327404Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-07-17T08:34:51.5330712Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-07-17T08:34:51.5335291Z * [new branch] gh/yanbing-j/34/base -> origin/gh/yanbing-j/34/base 2025-07-17T08:34:51.5338645Z * [new branch] gh/yanbing-j/34/head -> origin/gh/yanbing-j/34/head 2025-07-17T08:34:51.5342098Z * [new branch] gh/yanbing-j/34/orig -> origin/gh/yanbing-j/34/orig 2025-07-17T08:34:51.5346748Z * [new branch] gh/yanbing-j/36/base -> origin/gh/yanbing-j/36/base 2025-07-17T08:34:51.5350112Z * [new branch] gh/yanbing-j/36/head -> origin/gh/yanbing-j/36/head 2025-07-17T08:34:51.5353453Z * [new branch] gh/yanbing-j/36/orig -> origin/gh/yanbing-j/36/orig 2025-07-17T08:34:51.5361254Z * [new branch] gh/yanbing-j/37/base -> origin/gh/yanbing-j/37/base 2025-07-17T08:34:51.5364547Z * [new branch] gh/yanbing-j/37/head -> origin/gh/yanbing-j/37/head 2025-07-17T08:34:51.5367881Z * [new branch] gh/yanbing-j/37/orig -> origin/gh/yanbing-j/37/orig 2025-07-17T08:34:51.5372555Z * [new branch] gh/yanbing-j/38/base -> origin/gh/yanbing-j/38/base 2025-07-17T08:34:51.5376133Z * [new branch] gh/yanbing-j/38/head -> origin/gh/yanbing-j/38/head 2025-07-17T08:34:51.5379261Z * [new branch] gh/yanbing-j/38/orig -> origin/gh/yanbing-j/38/orig 2025-07-17T08:34:51.5383792Z * [new branch] gh/yanbing-j/39/base -> origin/gh/yanbing-j/39/base 2025-07-17T08:34:51.5387157Z * [new branch] gh/yanbing-j/39/head -> origin/gh/yanbing-j/39/head 2025-07-17T08:34:51.5390495Z * [new branch] gh/yanbing-j/39/orig -> origin/gh/yanbing-j/39/orig 2025-07-17T08:34:51.5396266Z * [new branch] gh/ydwu4/223/base -> origin/gh/ydwu4/223/base 2025-07-17T08:34:51.5399610Z * [new branch] gh/ydwu4/223/head -> origin/gh/ydwu4/223/head 2025-07-17T08:34:51.5403032Z * [new branch] gh/ydwu4/223/orig -> origin/gh/ydwu4/223/orig 2025-07-17T08:34:51.5407507Z * [new branch] gh/ydwu4/224/base -> origin/gh/ydwu4/224/base 2025-07-17T08:34:51.5410826Z * [new branch] gh/ydwu4/224/head -> origin/gh/ydwu4/224/head 2025-07-17T08:34:51.5414171Z * [new branch] gh/ydwu4/224/orig -> origin/gh/ydwu4/224/orig 2025-07-17T08:34:51.5418556Z * [new branch] gh/ydwu4/225/base -> origin/gh/ydwu4/225/base 2025-07-17T08:34:51.5421849Z * [new branch] gh/ydwu4/225/head -> origin/gh/ydwu4/225/head 2025-07-17T08:34:51.5425279Z * [new branch] gh/ydwu4/225/orig -> origin/gh/ydwu4/225/orig 2025-07-17T08:34:51.5429823Z * [new branch] gh/ydwu4/233/base -> origin/gh/ydwu4/233/base 2025-07-17T08:34:51.5433162Z * [new branch] gh/ydwu4/233/head -> origin/gh/ydwu4/233/head 2025-07-17T08:34:51.5437010Z * [new branch] gh/ydwu4/233/orig -> origin/gh/ydwu4/233/orig 2025-07-17T08:34:51.5441723Z * [new branch] gh/ydwu4/246/base -> origin/gh/ydwu4/246/base 2025-07-17T08:34:51.5445042Z * [new branch] gh/ydwu4/246/head -> origin/gh/ydwu4/246/head 2025-07-17T08:34:51.5448577Z * [new branch] gh/ydwu4/246/orig -> origin/gh/ydwu4/246/orig 2025-07-17T08:34:51.5453197Z * [new branch] gh/ydwu4/247/base -> origin/gh/ydwu4/247/base 2025-07-17T08:34:51.5456654Z * [new branch] gh/ydwu4/247/head -> origin/gh/ydwu4/247/head 2025-07-17T08:34:51.5459998Z * [new branch] gh/ydwu4/247/orig -> origin/gh/ydwu4/247/orig 2025-07-17T08:34:51.5464503Z * [new branch] gh/ydwu4/253/base -> origin/gh/ydwu4/253/base 2025-07-17T08:34:51.5467892Z * [new branch] gh/ydwu4/253/head -> origin/gh/ydwu4/253/head 2025-07-17T08:34:51.5471265Z * [new branch] gh/ydwu4/253/orig -> origin/gh/ydwu4/253/orig 2025-07-17T08:34:51.5476214Z * [new branch] gh/ydwu4/255/base -> origin/gh/ydwu4/255/base 2025-07-17T08:34:51.5479471Z * [new branch] gh/ydwu4/255/head -> origin/gh/ydwu4/255/head 2025-07-17T08:34:51.5482772Z * [new branch] gh/ydwu4/255/orig -> origin/gh/ydwu4/255/orig 2025-07-17T08:34:51.5487392Z * [new branch] gh/ydwu4/259/base -> origin/gh/ydwu4/259/base 2025-07-17T08:34:51.5490737Z * [new branch] gh/ydwu4/259/head -> origin/gh/ydwu4/259/head 2025-07-17T08:34:51.5494155Z * [new branch] gh/ydwu4/259/orig -> origin/gh/ydwu4/259/orig 2025-07-17T08:34:51.5498799Z * [new branch] gh/ydwu4/261/base -> origin/gh/ydwu4/261/base 2025-07-17T08:34:51.5502195Z * [new branch] gh/ydwu4/261/head -> origin/gh/ydwu4/261/head 2025-07-17T08:34:51.5505500Z * [new branch] gh/ydwu4/261/orig -> origin/gh/ydwu4/261/orig 2025-07-17T08:34:51.5509984Z * [new branch] gh/ydwu4/262/base -> origin/gh/ydwu4/262/base 2025-07-17T08:34:51.5513488Z * [new branch] gh/ydwu4/262/head -> origin/gh/ydwu4/262/head 2025-07-17T08:34:51.5516943Z * [new branch] gh/ydwu4/262/orig -> origin/gh/ydwu4/262/orig 2025-07-17T08:34:51.5521393Z * [new branch] gh/ydwu4/263/base -> origin/gh/ydwu4/263/base 2025-07-17T08:34:51.5524701Z * [new branch] gh/ydwu4/263/head -> origin/gh/ydwu4/263/head 2025-07-17T08:34:51.5528053Z * [new branch] gh/ydwu4/263/orig -> origin/gh/ydwu4/263/orig 2025-07-17T08:34:51.5532697Z * [new branch] gh/ydwu4/264/base -> origin/gh/ydwu4/264/base 2025-07-17T08:34:51.5536025Z * [new branch] gh/ydwu4/264/head -> origin/gh/ydwu4/264/head 2025-07-17T08:34:51.5539484Z * [new branch] gh/ydwu4/264/orig -> origin/gh/ydwu4/264/orig 2025-07-17T08:34:51.5544031Z * [new branch] gh/ydwu4/265/base -> origin/gh/ydwu4/265/base 2025-07-17T08:34:51.5547392Z * [new branch] gh/ydwu4/265/head -> origin/gh/ydwu4/265/head 2025-07-17T08:34:51.5550717Z * [new branch] gh/ydwu4/265/orig -> origin/gh/ydwu4/265/orig 2025-07-17T08:34:51.5555509Z * [new branch] gh/ydwu4/266/base -> origin/gh/ydwu4/266/base 2025-07-17T08:34:51.5558887Z * [new branch] gh/ydwu4/266/head -> origin/gh/ydwu4/266/head 2025-07-17T08:34:51.5562233Z * [new branch] gh/ydwu4/266/orig -> origin/gh/ydwu4/266/orig 2025-07-17T08:34:51.5566944Z * [new branch] gh/ydwu4/267/base -> origin/gh/ydwu4/267/base 2025-07-17T08:34:51.5570293Z * [new branch] gh/ydwu4/267/head -> origin/gh/ydwu4/267/head 2025-07-17T08:34:51.5573673Z * [new branch] gh/ydwu4/267/orig -> origin/gh/ydwu4/267/orig 2025-07-17T08:34:51.5578298Z * [new branch] gh/ydwu4/268/base -> origin/gh/ydwu4/268/base 2025-07-17T08:34:51.5581605Z * [new branch] gh/ydwu4/268/head -> origin/gh/ydwu4/268/head 2025-07-17T08:34:51.5585083Z * [new branch] gh/ydwu4/268/orig -> origin/gh/ydwu4/268/orig 2025-07-17T08:34:51.5589687Z * [new branch] gh/ydwu4/269/base -> origin/gh/ydwu4/269/base 2025-07-17T08:34:51.5592968Z * [new branch] gh/ydwu4/269/head -> origin/gh/ydwu4/269/head 2025-07-17T08:34:51.5596579Z * [new branch] gh/ydwu4/269/orig -> origin/gh/ydwu4/269/orig 2025-07-17T08:34:51.5601116Z * [new branch] gh/ydwu4/270/base -> origin/gh/ydwu4/270/base 2025-07-17T08:34:51.5604473Z * [new branch] gh/ydwu4/270/head -> origin/gh/ydwu4/270/head 2025-07-17T08:34:51.5607881Z * [new branch] gh/ydwu4/270/orig -> origin/gh/ydwu4/270/orig 2025-07-17T08:34:51.5612454Z * [new branch] gh/ydwu4/271/base -> origin/gh/ydwu4/271/base 2025-07-17T08:34:51.5615800Z * [new branch] gh/ydwu4/271/head -> origin/gh/ydwu4/271/head 2025-07-17T08:34:51.5619101Z * [new branch] gh/ydwu4/271/orig -> origin/gh/ydwu4/271/orig 2025-07-17T08:34:51.5623681Z * [new branch] gh/ydwu4/272/base -> origin/gh/ydwu4/272/base 2025-07-17T08:34:51.5627131Z * [new branch] gh/ydwu4/272/head -> origin/gh/ydwu4/272/head 2025-07-17T08:34:51.5630617Z * [new branch] gh/ydwu4/272/orig -> origin/gh/ydwu4/272/orig 2025-07-17T08:34:51.5635406Z * [new branch] gh/ydwu4/273/base -> origin/gh/ydwu4/273/base 2025-07-17T08:34:51.5638688Z * [new branch] gh/ydwu4/273/head -> origin/gh/ydwu4/273/head 2025-07-17T08:34:51.5642182Z * [new branch] gh/ydwu4/273/orig -> origin/gh/ydwu4/273/orig 2025-07-17T08:34:51.5646741Z * [new branch] gh/ydwu4/274/base -> origin/gh/ydwu4/274/base 2025-07-17T08:34:51.5650235Z * [new branch] gh/ydwu4/274/head -> origin/gh/ydwu4/274/head 2025-07-17T08:34:51.5653381Z * [new branch] gh/ydwu4/274/orig -> origin/gh/ydwu4/274/orig 2025-07-17T08:34:51.5657882Z * [new branch] gh/ydwu4/275/base -> origin/gh/ydwu4/275/base 2025-07-17T08:34:51.5661204Z * [new branch] gh/ydwu4/275/head -> origin/gh/ydwu4/275/head 2025-07-17T08:34:51.5664478Z * [new branch] gh/ydwu4/275/orig -> origin/gh/ydwu4/275/orig 2025-07-17T08:34:51.5668901Z * [new branch] gh/ydwu4/276/base -> origin/gh/ydwu4/276/base 2025-07-17T08:34:51.5672271Z * [new branch] gh/ydwu4/276/head -> origin/gh/ydwu4/276/head 2025-07-17T08:34:51.5675893Z * [new branch] gh/ydwu4/276/orig -> origin/gh/ydwu4/276/orig 2025-07-17T08:34:51.5680539Z * [new branch] gh/ydwu4/277/base -> origin/gh/ydwu4/277/base 2025-07-17T08:34:51.5683832Z * [new branch] gh/ydwu4/277/head -> origin/gh/ydwu4/277/head 2025-07-17T08:34:51.5687231Z * [new branch] gh/ydwu4/277/orig -> origin/gh/ydwu4/277/orig 2025-07-17T08:34:51.5691717Z * [new branch] gh/ydwu4/278/base -> origin/gh/ydwu4/278/base 2025-07-17T08:34:51.5695086Z * [new branch] gh/ydwu4/278/head -> origin/gh/ydwu4/278/head 2025-07-17T08:34:51.5698445Z * [new branch] gh/ydwu4/278/orig -> origin/gh/ydwu4/278/orig 2025-07-17T08:34:51.5702997Z * [new branch] gh/ydwu4/279/base -> origin/gh/ydwu4/279/base 2025-07-17T08:34:51.5706408Z * [new branch] gh/ydwu4/279/head -> origin/gh/ydwu4/279/head 2025-07-17T08:34:51.5709722Z * [new branch] gh/ydwu4/279/orig -> origin/gh/ydwu4/279/orig 2025-07-17T08:34:51.5716207Z * [new branch] gh/ydwu4/280/base -> origin/gh/ydwu4/280/base 2025-07-17T08:34:51.5720161Z * [new branch] gh/ydwu4/280/head -> origin/gh/ydwu4/280/head 2025-07-17T08:34:51.5723641Z * [new branch] gh/ydwu4/280/orig -> origin/gh/ydwu4/280/orig 2025-07-17T08:34:51.5728591Z * [new branch] gh/ydwu4/281/base -> origin/gh/ydwu4/281/base 2025-07-17T08:34:51.5731999Z * [new branch] gh/ydwu4/281/head -> origin/gh/ydwu4/281/head 2025-07-17T08:34:51.5735389Z * [new branch] gh/ydwu4/281/orig -> origin/gh/ydwu4/281/orig 2025-07-17T08:34:51.5739789Z * [new branch] gh/ydwu4/282/base -> origin/gh/ydwu4/282/base 2025-07-17T08:34:51.5743217Z * [new branch] gh/ydwu4/282/head -> origin/gh/ydwu4/282/head 2025-07-17T08:34:51.5746678Z * [new branch] gh/ydwu4/282/orig -> origin/gh/ydwu4/282/orig 2025-07-17T08:34:51.5751370Z * [new branch] gh/ydwu4/283/base -> origin/gh/ydwu4/283/base 2025-07-17T08:34:51.5754934Z * [new branch] gh/ydwu4/283/head -> origin/gh/ydwu4/283/head 2025-07-17T08:34:51.5758343Z * [new branch] gh/ydwu4/283/orig -> origin/gh/ydwu4/283/orig 2025-07-17T08:34:51.5763914Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-07-17T08:34:51.5767307Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-07-17T08:34:51.5772031Z * [new branch] gh/yf225/171/base -> origin/gh/yf225/171/base 2025-07-17T08:34:51.5775523Z * [new branch] gh/yf225/171/head -> origin/gh/yf225/171/head 2025-07-17T08:34:51.5778916Z * [new branch] gh/yf225/171/orig -> origin/gh/yf225/171/orig 2025-07-17T08:34:51.5783526Z * [new branch] gh/yf225/172/base -> origin/gh/yf225/172/base 2025-07-17T08:34:51.5786909Z * [new branch] gh/yf225/172/head -> origin/gh/yf225/172/head 2025-07-17T08:34:51.5790444Z * [new branch] gh/yf225/172/orig -> origin/gh/yf225/172/orig 2025-07-17T08:34:51.5795146Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-07-17T08:34:51.5798562Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-07-17T08:34:51.5804715Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-07-17T08:34:51.5808409Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-07-17T08:34:51.5811854Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-07-17T08:34:51.5816376Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-07-17T08:34:51.5819779Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-07-17T08:34:51.5823119Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-07-17T08:34:51.5828834Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-07-17T08:34:51.5832236Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-07-17T08:34:51.5837015Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-07-17T08:34:51.5840502Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-07-17T08:34:51.5846010Z * [new branch] gh/ysiraichi/79/base -> origin/gh/ysiraichi/79/base 2025-07-17T08:34:51.5849440Z * [new branch] gh/ysiraichi/79/head -> origin/gh/ysiraichi/79/head 2025-07-17T08:34:51.5852980Z * [new branch] gh/ysiraichi/79/orig -> origin/gh/ysiraichi/79/orig 2025-07-17T08:34:51.5857775Z * [new branch] gh/ysiraichi/80/base -> origin/gh/ysiraichi/80/base 2025-07-17T08:34:51.5861059Z * [new branch] gh/ysiraichi/80/head -> origin/gh/ysiraichi/80/head 2025-07-17T08:34:51.5864546Z * [new branch] gh/ysiraichi/80/orig -> origin/gh/ysiraichi/80/orig 2025-07-17T08:34:51.5869095Z * [new branch] gh/ysiraichi/81/base -> origin/gh/ysiraichi/81/base 2025-07-17T08:34:51.5872393Z * [new branch] gh/ysiraichi/81/head -> origin/gh/ysiraichi/81/head 2025-07-17T08:34:51.5876045Z * [new branch] gh/ysiraichi/81/orig -> origin/gh/ysiraichi/81/orig 2025-07-17T08:34:51.5880724Z * [new branch] gh/ysiraichi/84/base -> origin/gh/ysiraichi/84/base 2025-07-17T08:34:51.5884111Z * [new branch] gh/ysiraichi/84/head -> origin/gh/ysiraichi/84/head 2025-07-17T08:34:51.5887482Z * [new branch] gh/ysiraichi/84/orig -> origin/gh/ysiraichi/84/orig 2025-07-17T08:34:51.5892175Z * [new branch] gh/ysiraichi/85/base -> origin/gh/ysiraichi/85/base 2025-07-17T08:34:51.5895598Z * [new branch] gh/ysiraichi/85/head -> origin/gh/ysiraichi/85/head 2025-07-17T08:34:51.5899012Z * [new branch] gh/ysiraichi/85/orig -> origin/gh/ysiraichi/85/orig 2025-07-17T08:34:51.5903707Z * [new branch] gh/ysiraichi/86/base -> origin/gh/ysiraichi/86/base 2025-07-17T08:34:51.5907103Z * [new branch] gh/ysiraichi/86/head -> origin/gh/ysiraichi/86/head 2025-07-17T08:34:51.5910587Z * [new branch] gh/ysiraichi/86/orig -> origin/gh/ysiraichi/86/orig 2025-07-17T08:34:51.5915505Z * [new branch] gh/ysiraichi/87/base -> origin/gh/ysiraichi/87/base 2025-07-17T08:34:51.5918781Z * [new branch] gh/ysiraichi/87/head -> origin/gh/ysiraichi/87/head 2025-07-17T08:34:51.5922169Z * [new branch] gh/ysiraichi/87/orig -> origin/gh/ysiraichi/87/orig 2025-07-17T08:34:51.5927057Z * [new branch] gh/ysiraichi/88/base -> origin/gh/ysiraichi/88/base 2025-07-17T08:34:51.5930565Z * [new branch] gh/ysiraichi/88/head -> origin/gh/ysiraichi/88/head 2025-07-17T08:34:51.5933755Z * [new branch] gh/ysiraichi/88/orig -> origin/gh/ysiraichi/88/orig 2025-07-17T08:34:51.5939356Z * [new branch] gh/yuguo68/1/base -> origin/gh/yuguo68/1/base 2025-07-17T08:34:51.5942733Z * [new branch] gh/yuguo68/1/head -> origin/gh/yuguo68/1/head 2025-07-17T08:34:51.5946120Z * [new branch] gh/yuguo68/1/orig -> origin/gh/yuguo68/1/orig 2025-07-17T08:34:51.5950651Z * [new branch] gh/yuguo68/2/base -> origin/gh/yuguo68/2/base 2025-07-17T08:34:51.5954045Z * [new branch] gh/yuguo68/2/head -> origin/gh/yuguo68/2/head 2025-07-17T08:34:51.5957575Z * [new branch] gh/yuguo68/2/orig -> origin/gh/yuguo68/2/orig 2025-07-17T08:34:51.5963141Z * [new branch] gh/zhuhaozhe/28/base -> origin/gh/zhuhaozhe/28/base 2025-07-17T08:34:51.5966640Z * [new branch] gh/zhuhaozhe/28/head -> origin/gh/zhuhaozhe/28/head 2025-07-17T08:34:51.5969998Z * [new branch] gh/zhuhaozhe/28/orig -> origin/gh/zhuhaozhe/28/orig 2025-07-17T08:34:51.5974563Z * [new branch] gh/zhuhaozhe/29/base -> origin/gh/zhuhaozhe/29/base 2025-07-17T08:34:51.5977873Z * [new branch] gh/zhuhaozhe/29/head -> origin/gh/zhuhaozhe/29/head 2025-07-17T08:34:51.5981258Z * [new branch] gh/zhuhaozhe/29/orig -> origin/gh/zhuhaozhe/29/orig 2025-07-17T08:34:51.5985843Z * [new branch] gh/zhuhaozhe/31/base -> origin/gh/zhuhaozhe/31/base 2025-07-17T08:34:51.5989195Z * [new branch] gh/zhuhaozhe/31/head -> origin/gh/zhuhaozhe/31/head 2025-07-17T08:34:51.5992581Z * [new branch] gh/zhuhaozhe/31/orig -> origin/gh/zhuhaozhe/31/orig 2025-07-17T08:34:51.5997443Z * [new branch] gh/zhuhaozhe/32/base -> origin/gh/zhuhaozhe/32/base 2025-07-17T08:34:51.6000719Z * [new branch] gh/zhuhaozhe/32/head -> origin/gh/zhuhaozhe/32/head 2025-07-17T08:34:51.6004092Z * [new branch] gh/zhuhaozhe/32/orig -> origin/gh/zhuhaozhe/32/orig 2025-07-17T08:34:51.6008712Z * [new branch] gh/zhuhaozhe/33/base -> origin/gh/zhuhaozhe/33/base 2025-07-17T08:34:51.6012202Z * [new branch] gh/zhuhaozhe/33/head -> origin/gh/zhuhaozhe/33/head 2025-07-17T08:34:51.6015623Z * [new branch] gh/zhuhaozhe/33/orig -> origin/gh/zhuhaozhe/33/orig 2025-07-17T08:34:51.6021263Z * [new branch] gh/zhxchen17/22/base -> origin/gh/zhxchen17/22/base 2025-07-17T08:34:51.6024752Z * [new branch] gh/zhxchen17/22/head -> origin/gh/zhxchen17/22/head 2025-07-17T08:34:51.6028154Z * [new branch] gh/zhxchen17/22/orig -> origin/gh/zhxchen17/22/orig 2025-07-17T08:34:51.6032822Z * [new branch] gh/zhxchen17/25/base -> origin/gh/zhxchen17/25/base 2025-07-17T08:34:51.6036595Z * [new branch] gh/zhxchen17/25/head -> origin/gh/zhxchen17/25/head 2025-07-17T08:34:51.6039927Z * [new branch] gh/zhxchen17/25/orig -> origin/gh/zhxchen17/25/orig 2025-07-17T08:34:51.6044719Z * [new branch] gh/zhxchen17/26/base -> origin/gh/zhxchen17/26/base 2025-07-17T08:34:51.6048180Z * [new branch] gh/zhxchen17/26/head -> origin/gh/zhxchen17/26/head 2025-07-17T08:34:51.6051635Z * [new branch] gh/zhxchen17/26/orig -> origin/gh/zhxchen17/26/orig 2025-07-17T08:34:51.6056376Z * [new branch] gh/zhxchen17/27/base -> origin/gh/zhxchen17/27/base 2025-07-17T08:34:51.6059724Z * [new branch] gh/zhxchen17/27/head -> origin/gh/zhxchen17/27/head 2025-07-17T08:34:51.6063241Z * [new branch] gh/zhxchen17/27/orig -> origin/gh/zhxchen17/27/orig 2025-07-17T08:34:51.6069384Z * [new branch] gh/zhxchen17/28/base -> origin/gh/zhxchen17/28/base 2025-07-17T08:34:51.6071314Z * [new branch] gh/zhxchen17/28/head -> origin/gh/zhxchen17/28/head 2025-07-17T08:34:51.6074814Z * [new branch] gh/zhxchen17/28/orig -> origin/gh/zhxchen17/28/orig 2025-07-17T08:34:51.6079538Z * [new branch] gh/zhxchen17/29/base -> origin/gh/zhxchen17/29/base 2025-07-17T08:34:51.6082958Z * [new branch] gh/zhxchen17/29/head -> origin/gh/zhxchen17/29/head 2025-07-17T08:34:51.6086345Z * [new branch] gh/zhxchen17/29/orig -> origin/gh/zhxchen17/29/orig 2025-07-17T08:34:51.6090926Z * [new branch] gh/zhxchen17/30/base -> origin/gh/zhxchen17/30/base 2025-07-17T08:34:51.6094329Z * [new branch] gh/zhxchen17/30/head -> origin/gh/zhxchen17/30/head 2025-07-17T08:34:51.6097749Z * [new branch] gh/zhxchen17/30/orig -> origin/gh/zhxchen17/30/orig 2025-07-17T08:34:51.6102417Z * [new branch] gh/zhxchen17/31/base -> origin/gh/zhxchen17/31/base 2025-07-17T08:34:51.6105781Z * [new branch] gh/zhxchen17/31/head -> origin/gh/zhxchen17/31/head 2025-07-17T08:34:51.6109248Z * [new branch] gh/zhxchen17/31/orig -> origin/gh/zhxchen17/31/orig 2025-07-17T08:34:51.6113844Z * [new branch] gh/zhxchen17/32/base -> origin/gh/zhxchen17/32/base 2025-07-17T08:34:51.6117487Z * [new branch] gh/zhxchen17/32/head -> origin/gh/zhxchen17/32/head 2025-07-17T08:34:51.6120793Z * [new branch] gh/zhxchen17/32/orig -> origin/gh/zhxchen17/32/orig 2025-07-17T08:34:51.6126318Z * [new branch] gh/zklaus/1/base -> origin/gh/zklaus/1/base 2025-07-17T08:34:51.6129641Z * [new branch] gh/zklaus/1/head -> origin/gh/zklaus/1/head 2025-07-17T08:34:51.6132975Z * [new branch] gh/zklaus/1/orig -> origin/gh/zklaus/1/orig 2025-07-17T08:34:51.6137603Z * [new branch] gh/zklaus/10/base -> origin/gh/zklaus/10/base 2025-07-17T08:34:51.6141032Z * [new branch] gh/zklaus/10/head -> origin/gh/zklaus/10/head 2025-07-17T08:34:51.6144407Z * [new branch] gh/zklaus/10/orig -> origin/gh/zklaus/10/orig 2025-07-17T08:34:51.6148967Z * [new branch] gh/zklaus/11/base -> origin/gh/zklaus/11/base 2025-07-17T08:34:51.6152347Z * [new branch] gh/zklaus/11/head -> origin/gh/zklaus/11/head 2025-07-17T08:34:51.6155951Z * [new branch] gh/zklaus/11/orig -> origin/gh/zklaus/11/orig 2025-07-17T08:34:51.6160448Z * [new branch] gh/zklaus/12/base -> origin/gh/zklaus/12/base 2025-07-17T08:34:51.6163763Z * [new branch] gh/zklaus/12/head -> origin/gh/zklaus/12/head 2025-07-17T08:34:51.6167120Z * [new branch] gh/zklaus/12/orig -> origin/gh/zklaus/12/orig 2025-07-17T08:34:51.6171676Z * [new branch] gh/zklaus/13/base -> origin/gh/zklaus/13/base 2025-07-17T08:34:51.6175017Z * [new branch] gh/zklaus/13/head -> origin/gh/zklaus/13/head 2025-07-17T08:34:51.6178360Z * [new branch] gh/zklaus/13/orig -> origin/gh/zklaus/13/orig 2025-07-17T08:34:51.6182984Z * [new branch] gh/zklaus/14/base -> origin/gh/zklaus/14/base 2025-07-17T08:34:51.6186312Z * [new branch] gh/zklaus/14/head -> origin/gh/zklaus/14/head 2025-07-17T08:34:51.6189663Z * [new branch] gh/zklaus/14/orig -> origin/gh/zklaus/14/orig 2025-07-17T08:34:51.6194346Z * [new branch] gh/zklaus/15/base -> origin/gh/zklaus/15/base 2025-07-17T08:34:51.6197754Z * [new branch] gh/zklaus/15/head -> origin/gh/zklaus/15/head 2025-07-17T08:34:51.6201117Z * [new branch] gh/zklaus/15/orig -> origin/gh/zklaus/15/orig 2025-07-17T08:34:51.6206005Z * [new branch] gh/zklaus/16/base -> origin/gh/zklaus/16/base 2025-07-17T08:34:51.6209167Z * [new branch] gh/zklaus/16/head -> origin/gh/zklaus/16/head 2025-07-17T08:34:51.6212592Z * [new branch] gh/zklaus/16/orig -> origin/gh/zklaus/16/orig 2025-07-17T08:34:51.6217234Z * [new branch] gh/zklaus/17/base -> origin/gh/zklaus/17/base 2025-07-17T08:34:51.6220605Z * [new branch] gh/zklaus/17/head -> origin/gh/zklaus/17/head 2025-07-17T08:34:51.6223988Z * [new branch] gh/zklaus/17/orig -> origin/gh/zklaus/17/orig 2025-07-17T08:34:51.6228698Z * [new branch] gh/zklaus/18/base -> origin/gh/zklaus/18/base 2025-07-17T08:34:51.6232087Z * [new branch] gh/zklaus/18/head -> origin/gh/zklaus/18/head 2025-07-17T08:34:51.6235693Z * [new branch] gh/zklaus/18/orig -> origin/gh/zklaus/18/orig 2025-07-17T08:34:51.6240374Z * [new branch] gh/zklaus/2/base -> origin/gh/zklaus/2/base 2025-07-17T08:34:51.6243797Z * [new branch] gh/zklaus/2/head -> origin/gh/zklaus/2/head 2025-07-17T08:34:51.6247082Z * [new branch] gh/zklaus/2/orig -> origin/gh/zklaus/2/orig 2025-07-17T08:34:51.6251666Z * [new branch] gh/zklaus/4/base -> origin/gh/zklaus/4/base 2025-07-17T08:34:51.6255005Z * [new branch] gh/zklaus/4/head -> origin/gh/zklaus/4/head 2025-07-17T08:34:51.6258450Z * [new branch] gh/zklaus/4/orig -> origin/gh/zklaus/4/orig 2025-07-17T08:34:51.6263038Z * [new branch] gh/zklaus/7/base -> origin/gh/zklaus/7/base 2025-07-17T08:34:51.6266414Z * [new branch] gh/zklaus/7/head -> origin/gh/zklaus/7/head 2025-07-17T08:34:51.6269770Z * [new branch] gh/zklaus/7/orig -> origin/gh/zklaus/7/orig 2025-07-17T08:34:51.6274482Z * [new branch] gh/zklaus/8/base -> origin/gh/zklaus/8/base 2025-07-17T08:34:51.6277941Z * [new branch] gh/zklaus/8/head -> origin/gh/zklaus/8/head 2025-07-17T08:34:51.6281316Z * [new branch] gh/zklaus/8/orig -> origin/gh/zklaus/8/orig 2025-07-17T08:34:51.6285974Z * [new branch] gh/zklaus/9/base -> origin/gh/zklaus/9/base 2025-07-17T08:34:51.6289346Z * [new branch] gh/zklaus/9/head -> origin/gh/zklaus/9/head 2025-07-17T08:34:51.6292660Z * [new branch] gh/zklaus/9/orig -> origin/gh/zklaus/9/orig 2025-07-17T08:34:51.6298419Z * [new branch] gh/zou3519/1162/base -> origin/gh/zou3519/1162/base 2025-07-17T08:34:51.6301697Z * [new branch] gh/zou3519/1162/head -> origin/gh/zou3519/1162/head 2025-07-17T08:34:51.6305106Z * [new branch] gh/zou3519/1162/orig -> origin/gh/zou3519/1162/orig 2025-07-17T08:34:51.6309780Z * [new branch] gh/zou3519/1173/base -> origin/gh/zou3519/1173/base 2025-07-17T08:34:51.6313237Z * [new branch] gh/zou3519/1173/head -> origin/gh/zou3519/1173/head 2025-07-17T08:34:51.6316894Z * [new branch] gh/zou3519/1173/orig -> origin/gh/zou3519/1173/orig 2025-07-17T08:34:51.6321413Z * [new branch] gh/zou3519/1174/base -> origin/gh/zou3519/1174/base 2025-07-17T08:34:51.6324821Z * [new branch] gh/zou3519/1174/head -> origin/gh/zou3519/1174/head 2025-07-17T08:34:51.6328314Z * [new branch] gh/zou3519/1174/orig -> origin/gh/zou3519/1174/orig 2025-07-17T08:34:51.6332854Z * [new branch] gh/zou3519/1175/base -> origin/gh/zou3519/1175/base 2025-07-17T08:34:51.6336225Z * [new branch] gh/zou3519/1175/head -> origin/gh/zou3519/1175/head 2025-07-17T08:34:51.6339663Z * [new branch] gh/zou3519/1175/orig -> origin/gh/zou3519/1175/orig 2025-07-17T08:34:51.6344246Z * [new branch] gh/zou3519/1176/base -> origin/gh/zou3519/1176/base 2025-07-17T08:34:51.6347382Z * [new branch] gh/zou3519/1176/head -> origin/gh/zou3519/1176/head 2025-07-17T08:34:51.6350836Z * [new branch] gh/zou3519/1176/orig -> origin/gh/zou3519/1176/orig 2025-07-17T08:34:51.6355434Z * [new branch] gh/zou3519/1177/base -> origin/gh/zou3519/1177/base 2025-07-17T08:34:51.6358806Z * [new branch] gh/zou3519/1177/head -> origin/gh/zou3519/1177/head 2025-07-17T08:34:51.6362151Z * [new branch] gh/zou3519/1177/orig -> origin/gh/zou3519/1177/orig 2025-07-17T08:34:51.6366899Z * [new branch] gh/zou3519/1178/base -> origin/gh/zou3519/1178/base 2025-07-17T08:34:51.6371451Z * [new branch] gh/zou3519/1178/head -> origin/gh/zou3519/1178/head 2025-07-17T08:34:51.6373754Z * [new branch] gh/zou3519/1178/orig -> origin/gh/zou3519/1178/orig 2025-07-17T08:34:51.6378534Z * [new branch] gh/zou3519/1179/base -> origin/gh/zou3519/1179/base 2025-07-17T08:34:51.6381833Z * [new branch] gh/zou3519/1179/head -> origin/gh/zou3519/1179/head 2025-07-17T08:34:51.6385219Z * [new branch] gh/zou3519/1179/orig -> origin/gh/zou3519/1179/orig 2025-07-17T08:34:51.6389637Z * [new branch] gh/zou3519/1180/base -> origin/gh/zou3519/1180/base 2025-07-17T08:34:51.6392982Z * [new branch] gh/zou3519/1180/head -> origin/gh/zou3519/1180/head 2025-07-17T08:34:51.6396633Z * [new branch] gh/zou3519/1180/orig -> origin/gh/zou3519/1180/orig 2025-07-17T08:34:51.6401019Z * [new branch] gh/zou3519/1181/base -> origin/gh/zou3519/1181/base 2025-07-17T08:34:51.6404380Z * [new branch] gh/zou3519/1181/head -> origin/gh/zou3519/1181/head 2025-07-17T08:34:51.6407810Z * [new branch] gh/zou3519/1181/orig -> origin/gh/zou3519/1181/orig 2025-07-17T08:34:51.6412402Z * [new branch] gh/zou3519/1182/base -> origin/gh/zou3519/1182/base 2025-07-17T08:34:51.6415775Z * [new branch] gh/zou3519/1182/head -> origin/gh/zou3519/1182/head 2025-07-17T08:34:51.6419418Z * [new branch] gh/zou3519/1182/orig -> origin/gh/zou3519/1182/orig 2025-07-17T08:34:51.6424071Z * [new branch] gh/zou3519/1183/base -> origin/gh/zou3519/1183/base 2025-07-17T08:34:51.6427510Z * [new branch] gh/zou3519/1183/head -> origin/gh/zou3519/1183/head 2025-07-17T08:34:51.6430838Z * [new branch] gh/zou3519/1183/orig -> origin/gh/zou3519/1183/orig 2025-07-17T08:34:51.6435763Z * [new branch] gh/zou3519/1184/base -> origin/gh/zou3519/1184/base 2025-07-17T08:34:51.6439097Z * [new branch] gh/zou3519/1184/head -> origin/gh/zou3519/1184/head 2025-07-17T08:34:51.6442537Z * [new branch] gh/zou3519/1184/orig -> origin/gh/zou3519/1184/orig 2025-07-17T08:34:51.6446933Z * [new branch] gh/zou3519/1185/base -> origin/gh/zou3519/1185/base 2025-07-17T08:34:51.6450317Z * [new branch] gh/zou3519/1185/head -> origin/gh/zou3519/1185/head 2025-07-17T08:34:51.6453786Z * [new branch] gh/zou3519/1185/orig -> origin/gh/zou3519/1185/orig 2025-07-17T08:34:51.6458217Z * [new branch] gh/zou3519/1186/base -> origin/gh/zou3519/1186/base 2025-07-17T08:34:51.6461529Z * [new branch] gh/zou3519/1186/head -> origin/gh/zou3519/1186/head 2025-07-17T08:34:51.6465017Z * [new branch] gh/zou3519/1186/orig -> origin/gh/zou3519/1186/orig 2025-07-17T08:34:51.6469724Z * [new branch] gh/zou3519/916/base -> origin/gh/zou3519/916/base 2025-07-17T08:34:51.6473132Z * [new branch] gh/zou3519/916/head -> origin/gh/zou3519/916/head 2025-07-17T08:34:51.6479445Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-07-17T08:34:51.6482373Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-07-17T08:34:51.6486998Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-07-17T08:34:51.6490400Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-07-17T08:34:51.6494842Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-07-17T08:34:51.6498121Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-07-17T08:34:51.6502527Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-07-17T08:34:51.6505935Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-07-17T08:34:51.6510373Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-07-17T08:34:51.6513749Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-07-17T08:34:51.6518359Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-07-17T08:34:51.6521644Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-07-17T08:34:51.6526055Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-07-17T08:34:51.6529437Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-07-17T08:34:51.6533174Z * [new branch] google-main -> origin/google-main 2025-07-17T08:34:51.6537677Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-07-17T08:34:51.6540838Z * [new branch] guangyey/host_alloc -> origin/guangyey/host_alloc 2025-07-17T08:34:51.6544090Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-07-17T08:34:51.6547373Z * [new branch] guard_system -> origin/guard_system 2025-07-17T08:34:51.6552069Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-07-17T08:34:51.6556587Z * [new branch] haozhe/bf16-dynamic-shape -> origin/haozhe/bf16-dynamic-shape 2025-07-17T08:34:51.6560017Z * [new branch] hc_baseline -> origin/hc_baseline 2025-07-17T08:34:51.6563276Z * [new branch] hhh_decomp_mul -> origin/hhh_decomp_mul 2025-07-17T08:34:51.6566524Z * [new branch] hhh_rand -> origin/hhh_rand 2025-07-17T08:34:51.6571145Z * [new branch] hoy/mmsplitk -> origin/hoy/mmsplitk 2025-07-17T08:34:51.6574462Z * [new branch] hoy/triton-PR3973 -> origin/hoy/triton-PR3973 2025-07-17T08:34:51.6577810Z * [new branch] hoy/triton-coalescing-baseline -> origin/hoy/triton-coalescing-baseline 2025-07-17T08:34:51.6581082Z * [new branch] hoy/triton-coalescing-min -> origin/hoy/triton-coalescing-min 2025-07-17T08:34:51.6584320Z * [new branch] hoy/triton-coalescing-new -> origin/hoy/triton-coalescing-new 2025-07-17T08:34:51.6587571Z * [new branch] hoy/triton-coalescing-vec -> origin/hoy/triton-coalescing-vec 2025-07-17T08:34:51.6590788Z * [new branch] inductordecompfix -> origin/inductordecompfix 2025-07-17T08:34:51.6594165Z * [new branch] inline -> origin/inline 2025-07-17T08:34:51.6597729Z * [new branch] inlining -> origin/inlining 2025-07-17T08:34:51.6601090Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-07-17T08:34:51.6604596Z * [new branch] int8_sdpa -> origin/int8_sdpa 2025-07-17T08:34:51.6608368Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-07-17T08:34:51.6611428Z * [new branch] is_contiguous_shim -> origin/is_contiguous_shim 2025-07-17T08:34:51.6614787Z * [new branch] issue#58739 -> origin/issue#58739 2025-07-17T08:34:51.6618088Z * [new branch] issue-154849 -> origin/issue-154849 2025-07-17T08:34:51.6622595Z * [new branch] ivanov/cherry-pick-ckpt-fixes -> origin/ivanov/cherry-pick-ckpt-fixes 2025-07-17T08:34:51.6627245Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-07-17T08:34:51.6630665Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-07-17T08:34:51.6639435Z * [new branch] justinchu/attention-tests -> origin/justinchu/attention-tests 2025-07-17T08:34:51.6642754Z * [new branch] justinchu/del-opset-tool -> origin/justinchu/del-opset-tool 2025-07-17T08:34:51.6645995Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-07-17T08:34:51.6650786Z * [new branch] leslie/enable_poc_reduction_fusion -> origin/leslie/enable_poc_reduction_fusion 2025-07-17T08:34:51.6654102Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-07-17T08:34:51.6658490Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-07-17T08:34:51.6662793Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-07-17T08:34:51.6666020Z * [new branch] liaoxuan/tags_issue -> origin/liaoxuan/tags_issue 2025-07-17T08:34:51.6669341Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-07-17T08:34:51.6672493Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-07-17T08:34:51.6675828Z * [new branch] local_map -> origin/local_map 2025-07-17T08:34:51.6679153Z * [new branch] logdetfix -> origin/logdetfix 2025-07-17T08:34:51.6684628Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-07-17T08:34:51.6689166Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-07-17T08:34:51.6692430Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-07-17T08:34:51.6695789Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-07-17T08:34:51.6699107Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-07-17T08:34:51.6702337Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-07-17T08:34:51.6705600Z * [new branch] lucaskabela/typing_bytecode_analysis_transform -> origin/lucaskabela/typing_bytecode_analysis_transform 2025-07-17T08:34:51.6708762Z * [new branch] lucaskabela/typing_cache_files -> origin/lucaskabela/typing_cache_files 2025-07-17T08:34:51.6712072Z * [new branch] lucaskabela/typing_debug_utils.py -> origin/lucaskabela/typing_debug_utils.py 2025-07-17T08:34:51.6715420Z * [new branch] lucaskabela/typing_decorators -> origin/lucaskabela/typing_decorators 2025-07-17T08:34:51.6718622Z * [new branch] lucaskabela/typing_eval_frame -> origin/lucaskabela/typing_eval_frame 2025-07-17T08:34:51.6721864Z * [new branch] lucaskabela/typing_for_codegen -> origin/lucaskabela/typing_for_codegen 2025-07-17T08:34:51.6725103Z * [new branch] lucaskabela/typing_source_guard -> origin/lucaskabela/typing_source_guard 2025-07-17T08:34:51.6729881Z * [new branch] m54qxk-codex/update-default-size_hint-fallback-value -> origin/m54qxk-codex/update-default-size_hint-fallback-value 2025-07-17T08:34:51.6733025Z * [new branch] main -> origin/main 2025-07-17T08:34:51.6736359Z * [new branch] malfet-patch-1 -> origin/malfet-patch-1 2025-07-17T08:34:51.6739658Z * [new branch] malfet-patch-10 -> origin/malfet-patch-10 2025-07-17T08:34:51.6742902Z * [new branch] malfet-patch-11 -> origin/malfet-patch-11 2025-07-17T08:34:51.6746264Z * [new branch] malfet-patch-12 -> origin/malfet-patch-12 2025-07-17T08:34:51.6749542Z * [new branch] malfet-patch-13 -> origin/malfet-patch-13 2025-07-17T08:34:51.6752728Z * [new branch] malfet-patch-14 -> origin/malfet-patch-14 2025-07-17T08:34:51.6756410Z * [new branch] malfet-patch-15 -> origin/malfet-patch-15 2025-07-17T08:34:51.6759603Z * [new branch] malfet-patch-16 -> origin/malfet-patch-16 2025-07-17T08:34:51.6762968Z * [new branch] malfet-patch-2 -> origin/malfet-patch-2 2025-07-17T08:34:51.6766339Z * [new branch] malfet-patch-21 -> origin/malfet-patch-21 2025-07-17T08:34:51.6769685Z * [new branch] malfet-patch-23 -> origin/malfet-patch-23 2025-07-17T08:34:51.6772917Z * [new branch] malfet-patch-3 -> origin/malfet-patch-3 2025-07-17T08:34:51.6776242Z * [new branch] malfet-patch-4 -> origin/malfet-patch-4 2025-07-17T08:34:51.6779496Z * [new branch] malfet-patch-5 -> origin/malfet-patch-5 2025-07-17T08:34:51.6782924Z * [new branch] malfet-patch-6 -> origin/malfet-patch-6 2025-07-17T08:34:51.6786290Z * [new branch] malfet-patch-7 -> origin/malfet-patch-7 2025-07-17T08:34:51.6789779Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-07-17T08:34:51.6793078Z * [new branch] malfet-patch-9 -> origin/malfet-patch-9 2025-07-17T08:34:51.6799889Z * [new branch] malfet/delete-upsteam-cuda -> origin/malfet/delete-upsteam-cuda 2025-07-17T08:34:51.6803100Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-07-17T08:34:51.6807478Z * [new branch] manuel/bench-scan-metal -> origin/manuel/bench-scan-metal 2025-07-17T08:34:51.6810768Z * [new branch] manuel/fix_multidim_boolean_indexing -> origin/manuel/fix_multidim_boolean_indexing 2025-07-17T08:34:51.6813928Z * [new branch] manuel/np_empty_ellipsis -> origin/manuel/np_empty_ellipsis 2025-07-17T08:34:51.6817153Z * [new branch] manuel/scan-metal -> origin/manuel/scan-metal 2025-07-17T08:34:51.6820380Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-07-17T08:34:51.6824811Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-07-17T08:34:51.6828076Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-07-17T08:34:51.6831377Z * [new branch] mlazos/backup-test-branch -> origin/mlazos/backup-test-branch 2025-07-17T08:34:51.6834731Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-07-17T08:34:51.6838074Z * [new branch] mlazos/baseline -> origin/mlazos/baseline 2025-07-17T08:34:51.6841416Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-07-17T08:34:51.6844655Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-07-17T08:34:51.6847966Z * [new branch] mlazos/buf-fix -> origin/mlazos/buf-fix 2025-07-17T08:34:51.6851182Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-07-17T08:34:51.6854541Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-07-17T08:34:51.6857501Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-07-17T08:34:51.6861048Z * [new branch] mlazos/ck2 -> origin/mlazos/ck2 2025-07-17T08:34:51.6864371Z * [new branch] mlazos/combokernels -> origin/mlazos/combokernels 2025-07-17T08:34:51.6867814Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-07-17T08:34:51.6871262Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-07-17T08:34:51.6874921Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-07-17T08:34:51.6878165Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-07-17T08:34:51.6881409Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-07-17T08:34:51.6884635Z * [new branch] mlazos/data-gather -> origin/mlazos/data-gather 2025-07-17T08:34:51.6887952Z * [new branch] mlazos/data-ptrs2 -> origin/mlazos/data-ptrs2 2025-07-17T08:34:51.6891086Z * [new branch] mlazos/data-ptrs3 -> origin/mlazos/data-ptrs3 2025-07-17T08:34:51.6894535Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-07-17T08:34:51.6898036Z * [new branch] mlazos/disable-closures -> origin/mlazos/disable-closures 2025-07-17T08:34:51.6901375Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-07-17T08:34:51.6904600Z * [new branch] mlazos/disabled-opt -> origin/mlazos/disabled-opt 2025-07-17T08:34:51.6907877Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-07-17T08:34:51.6911087Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-07-17T08:34:51.6914493Z * [new branch] mlazos/exp_disable -> origin/mlazos/exp_disable 2025-07-17T08:34:51.6917742Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-07-17T08:34:51.6921083Z * [new branch] mlazos/foreach-reds -> origin/mlazos/foreach-reds 2025-07-17T08:34:51.6924377Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-07-17T08:34:51.6927727Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-07-17T08:34:51.6930975Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-07-17T08:34:51.6934175Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-07-17T08:34:51.6937429Z * [new branch] mlazos/gen-foreach -> origin/mlazos/gen-foreach 2025-07-17T08:34:51.6940568Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-07-17T08:34:51.6943791Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-07-17T08:34:51.6947237Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-07-17T08:34:51.6950621Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-07-17T08:34:51.6953954Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-07-17T08:34:51.6957608Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-07-17T08:34:51.6961026Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-07-17T08:34:51.6964250Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-07-17T08:34:51.6967472Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-07-17T08:34:51.6970697Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-07-17T08:34:51.6974146Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-07-17T08:34:51.6977168Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-07-17T08:34:51.6980426Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-07-17T08:34:51.6983664Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-07-17T08:34:51.6986989Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-07-17T08:34:51.6990283Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-07-17T08:34:51.6993472Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-07-17T08:34:51.7003262Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-07-17T08:34:51.7003540Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-07-17T08:34:51.7003840Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-07-17T08:34:51.7006561Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-07-17T08:34:51.7009763Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-07-17T08:34:51.7013063Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-07-17T08:34:51.7016294Z * [new branch] mlazos/hop-modes -> origin/mlazos/hop-modes 2025-07-17T08:34:51.7019524Z * [new branch] mlazos/init-per-param -> origin/mlazos/init-per-param 2025-07-17T08:34:51.7022735Z * [new branch] mlazos/init_per_param -> origin/mlazos/init_per_param 2025-07-17T08:34:51.7026012Z * [new branch] mlazos/less-guards -> origin/mlazos/less-guards 2025-07-17T08:34:51.7029402Z * [new branch] mlazos/lr-composibility -> origin/mlazos/lr-composibility 2025-07-17T08:34:51.7032520Z * [new branch] mlazos/main -> origin/mlazos/main 2025-07-17T08:34:51.7035919Z * [new branch] mlazos/main-test-enablement -> origin/mlazos/main-test-enablement 2025-07-17T08:34:51.7039061Z * [new branch] mlazos/main2 -> origin/mlazos/main2 2025-07-17T08:34:51.7042323Z * [new branch] mlazos/mcg -> origin/mlazos/mcg 2025-07-17T08:34:51.7045572Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-07-17T08:34:51.7048813Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-07-17T08:34:51.7053283Z * [new branch] mlazos/mlazos/ck2 -> origin/mlazos/mlazos/ck2 2025-07-17T08:34:51.7056651Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-07-17T08:34:51.7060082Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-07-17T08:34:51.7063321Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-07-17T08:34:51.7066633Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-07-17T08:34:51.7069852Z * [new branch] mlazos/more-tests -> origin/mlazos/more-tests 2025-07-17T08:34:51.7073079Z * [new branch] mlazos/nested-dc -> origin/mlazos/nested-dc 2025-07-17T08:34:51.7076456Z * [new branch] mlazos/no-cpp -> origin/mlazos/no-cpp 2025-07-17T08:34:51.7079742Z * [new branch] mlazos/no-init-group-handling -> origin/mlazos/no-init-group-handling 2025-07-17T08:34:51.7083030Z * [new branch] mlazos/op-investigation -> origin/mlazos/op-investigation 2025-07-17T08:34:51.7086201Z * [new branch] mlazos/opt-bench-exp2 -> origin/mlazos/opt-bench-exp2 2025-07-17T08:34:51.7089466Z * [new branch] mlazos/opt-bench3 -> origin/mlazos/opt-bench3 2025-07-17T08:34:51.7093103Z * [new branch] mlazos/opt-incr -> origin/mlazos/opt-incr 2025-07-17T08:34:51.7095887Z * [new branch] mlazos/opt-slowdown -> origin/mlazos/opt-slowdown 2025-07-17T08:34:51.7099121Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-07-17T08:34:51.7102481Z * [new branch] mlazos/proxy-opt -> origin/mlazos/proxy-opt 2025-07-17T08:34:51.7105737Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-07-17T08:34:51.7108963Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-07-17T08:34:51.7112229Z * [new branch] mlazos/sdpa-driss -> origin/mlazos/sdpa-driss 2025-07-17T08:34:51.7115723Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-07-17T08:34:51.7119010Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-07-17T08:34:51.7122574Z * [new branch] mlazos/sub-param-fix -> origin/mlazos/sub-param-fix 2025-07-17T08:34:51.7126047Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-07-17T08:34:51.7129356Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-07-17T08:34:51.7132532Z * [new branch] mlazos/test -> origin/mlazos/test 2025-07-17T08:34:51.7135829Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-07-17T08:34:51.7139097Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-07-17T08:34:51.7142369Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-07-17T08:34:51.7145856Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-07-17T08:34:51.7149100Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-07-17T08:34:51.7152396Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-07-17T08:34:51.7155861Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-07-17T08:34:51.7159078Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-07-17T08:34:51.7162277Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-07-17T08:34:51.7165495Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-07-17T08:34:51.7168783Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-07-17T08:34:51.7172076Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-07-17T08:34:51.7175394Z * [new branch] move-theme-out-docker -> origin/move-theme-out-docker 2025-07-17T08:34:51.7178758Z * [new branch] mps-linear-1d -> origin/mps-linear-1d 2025-07-17T08:34:51.7183140Z * [new branch] msaroufim/acc -> origin/msaroufim/acc 2025-07-17T08:34:51.7186390Z * [new branch] msaroufim/be1 -> origin/msaroufim/be1 2025-07-17T08:34:51.7189643Z * [new branch] msaroufim/ck_custom_op -> origin/msaroufim/ck_custom_op 2025-07-17T08:34:51.7192984Z * [new branch] msaroufim/cn_path -> origin/msaroufim/cn_path 2025-07-17T08:34:51.7196450Z * [new branch] msaroufim/compile_kernel -> origin/msaroufim/compile_kernel 2025-07-17T08:34:51.7199641Z * [new branch] msaroufim/compile_reduction -> origin/msaroufim/compile_reduction 2025-07-17T08:34:51.7202947Z * [new branch] msaroufim/dtensorfusedadam -> origin/msaroufim/dtensorfusedadam 2025-07-17T08:34:51.7206297Z * [new branch] msaroufim/gencode -> origin/msaroufim/gencode 2025-07-17T08:34:51.7209655Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-07-17T08:34:51.7213140Z * [new branch] msaroufim/simplify_nvrtc_discovery -> origin/msaroufim/simplify_nvrtc_discovery 2025-07-17T08:34:51.7217226Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-07-17T08:34:51.7220436Z * [new branch] mul_op -> origin/mul_op 2025-07-17T08:34:51.7224822Z * [new branch] ngimel/cat_perf -> origin/ngimel/cat_perf 2025-07-17T08:34:51.7228098Z * [new branch] ngimel/deterministic_scatter -> origin/ngimel/deterministic_scatter 2025-07-17T08:34:51.7231242Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-07-17T08:34:51.7234643Z * [new branch] ngimel/index_None -> origin/ngimel/index_None 2025-07-17T08:34:51.7238039Z * [new branch] ngimel/mempool_emptycache -> origin/ngimel/mempool_emptycache 2025-07-17T08:34:51.7241193Z * [new branch] nightly -> origin/nightly 2025-07-17T08:34:51.7244651Z * [new branch] nmacchioni-be-use-max-autotune -> origin/nmacchioni-be-use-max-autotune 2025-07-17T08:34:51.7247988Z * [new branch] nmacchioni-patch-1 -> origin/nmacchioni-patch-1 2025-07-17T08:34:51.7251450Z * [new branch] nmacchioni-patch-10 -> origin/nmacchioni-patch-10 2025-07-17T08:34:51.7254801Z * [new branch] nmacchioni-patch-3 -> origin/nmacchioni-patch-3 2025-07-17T08:34:51.7258178Z * [new branch] nmacchioni-patch-4 -> origin/nmacchioni-patch-4 2025-07-17T08:34:51.7261404Z * [new branch] nmacchioni-patch-5 -> origin/nmacchioni-patch-5 2025-07-17T08:34:51.7264748Z * [new branch] nmacchioni-patch-6 -> origin/nmacchioni-patch-6 2025-07-17T08:34:51.7268010Z * [new branch] nmacchioni-patch-7 -> origin/nmacchioni-patch-7 2025-07-17T08:34:51.7271332Z * [new branch] nmacchioni-patch-8 -> origin/nmacchioni-patch-8 2025-07-17T08:34:51.7274735Z * [new branch] nmacchioni-patch-9 -> origin/nmacchioni-patch-9 2025-07-17T08:34:51.7278085Z * [new branch] np_indexing_fix -> origin/np_indexing_fix 2025-07-17T08:34:51.7281364Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-07-17T08:34:51.7285861Z * [new branch] nweidia/enable-B200-inductor-nightly-ci -> origin/nweidia/enable-B200-inductor-nightly-ci 2025-07-17T08:34:51.7289106Z * [new branch] openblas_gemv -> origin/openblas_gemv 2025-07-17T08:34:51.7294623Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-07-17T08:34:51.7298218Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-07-17T08:34:51.7301751Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-07-17T08:34:51.7305319Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-07-17T08:34:51.7308768Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-07-17T08:34:51.7312291Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-07-17T08:34:51.7315853Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-07-17T08:34:51.7319293Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-07-17T08:34:51.7322694Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-07-17T08:34:51.7326109Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-07-17T08:34:51.7329530Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-07-17T08:34:51.7332832Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-07-17T08:34:51.7336304Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-07-17T08:34:51.7339401Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-07-17T08:34:51.7342801Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-07-17T08:34:51.7346000Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-07-17T08:34:51.7349154Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-07-17T08:34:51.7353634Z * [new branch] oulgen/fx_graph -> origin/oulgen/fx_graph 2025-07-17T08:34:51.7356981Z * [new branch] padded-tensor -> origin/padded-tensor 2025-07-17T08:34:51.7360645Z * [new branch] parallel_cat -> origin/parallel_cat 2025-07-17T08:34:51.7363919Z * [new branch] parallel_reduce -> origin/parallel_reduce 2025-07-17T08:34:51.7367287Z * [new branch] pca2 -> origin/pca2 2025-07-17T08:34:51.7370668Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-07-17T08:34:51.7375190Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-07-17T08:34:51.7378347Z * [new branch] pianpwk/dde_repeat_cat -> origin/pianpwk/dde_repeat_cat 2025-07-17T08:34:51.7381607Z * [new branch] pianpwk/dim_docs -> origin/pianpwk/dim_docs 2025-07-17T08:34:51.7385004Z * [new branch] pianpwk/draft_export_normalize -> origin/pianpwk/draft_export_normalize 2025-07-17T08:34:51.7388247Z * [new branch] pianpwk/dynamic_source_dim -> origin/pianpwk/dynamic_source_dim 2025-07-17T08:34:51.7391471Z * [new branch] pianpwk/int_in_pgo_state -> origin/pianpwk/int_in_pgo_state 2025-07-17T08:34:51.7394887Z * [new branch] pianpwk/invalidate_fake_memo -> origin/pianpwk/invalidate_fake_memo 2025-07-17T08:34:51.7398265Z * [new branch] pianpwk/lru_cache_bound_sympy -> origin/pianpwk/lru_cache_bound_sympy 2025-07-17T08:34:51.7401448Z * [new branch] pianpwk/max_1_strides -> origin/pianpwk/max_1_strides 2025-07-17T08:34:51.7404765Z * [new branch] pianpwk/nonzero_memo -> origin/pianpwk/nonzero_memo 2025-07-17T08:34:51.7408102Z * [new branch] pianpwk/obl_scatter_gather_index -> origin/pianpwk/obl_scatter_gather_index 2025-07-17T08:34:51.7411315Z * [new branch] pianpwk/oblivious_infer_size -> origin/pianpwk/oblivious_infer_size 2025-07-17T08:34:51.7414622Z * [new branch] pianpwk/oblivious_meta_select -> origin/pianpwk/oblivious_meta_select 2025-07-17T08:34:51.7417899Z * [new branch] pianpwk/oblivious_reshape_view_bad -> origin/pianpwk/oblivious_reshape_view_bad 2025-07-17T08:34:51.7421167Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-07-17T08:34:51.7424417Z * [new branch] pianpwk/oblivious_should_swap -> origin/pianpwk/oblivious_should_swap 2025-07-17T08:34:51.7427684Z * [new branch] pianpwk/oblivious_slice_forward -> origin/pianpwk/oblivious_slice_forward 2025-07-17T08:34:51.7430829Z * [new branch] pianpwk/oblivious_where -> origin/pianpwk/oblivious_where 2025-07-17T08:34:51.7434092Z * [new branch] pianpwk/pad_nd_oblivious -> origin/pianpwk/pad_nd_oblivious 2025-07-17T08:34:51.7437389Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-07-17T08:34:51.7440697Z * [new branch] pianpwk/remove_guard_fail_break -> origin/pianpwk/remove_guard_fail_break 2025-07-17T08:34:51.7443953Z * [new branch] pianpwk/single_slice -> origin/pianpwk/single_slice 2025-07-17T08:34:51.7447345Z * [new branch] pianpwk/slice_fresh_symbols -> origin/pianpwk/slice_fresh_symbols 2025-07-17T08:34:51.7450796Z * [new branch] pianpwk/test_slice_fake_impl -> origin/pianpwk/test_slice_fake_impl 2025-07-17T08:34:51.7453895Z * [new branch] pianpwk/unbacked_channels_last -> origin/pianpwk/unbacked_channels_last 2025-07-17T08:34:51.7457340Z * [new branch] pianpwk/unbacked_safe_conv1d -> origin/pianpwk/unbacked_safe_conv1d 2025-07-17T08:34:51.7460609Z * [new branch] pianpwk/unbacked_sdpa_flash -> origin/pianpwk/unbacked_sdpa_flash 2025-07-17T08:34:51.7463891Z * [new branch] pianpwk/unbacked_should_swap -> origin/pianpwk/unbacked_should_swap 2025-07-17T08:34:51.7467365Z * [new branch] pianpwk/unbacked_slice_binding -> origin/pianpwk/unbacked_slice_binding 2025-07-17T08:34:51.7470600Z * [new branch] pianpwk/unbacked_slice_forward -> origin/pianpwk/unbacked_slice_forward 2025-07-17T08:34:51.7473983Z * [new branch] pianpwk/verbose_tensor_guards -> origin/pianpwk/verbose_tensor_guards 2025-07-17T08:34:51.7477365Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-07-17T08:34:51.7480647Z * [new branch] pianpwk/whitelist_int_float -> origin/pianpwk/whitelist_int_float 2025-07-17T08:34:51.7483863Z * [new branch] pianpwk/whitelist_optimizer -> origin/pianpwk/whitelist_optimizer 2025-07-17T08:34:51.7487351Z * [new branch] pick-torch-standalone-rename-to-release-2.8 -> origin/pick-torch-standalone-rename-to-release-2.8 2025-07-17T08:34:51.7490539Z * [new branch] pin-theme-version -> origin/pin-theme-version 2025-07-17T08:34:51.7493789Z * [new branch] pin-torchao -> origin/pin-torchao 2025-07-17T08:34:51.7498203Z * [new branch] piz/clean_strategy_0627 -> origin/piz/clean_strategy_0627 2025-07-17T08:34:51.7501470Z * [new branch] piz/dtensor_op_coverage -> origin/piz/dtensor_op_coverage 2025-07-17T08:34:51.7504743Z * [new branch] piz/fall_back_missing_0705 -> origin/piz/fall_back_missing_0705 2025-07-17T08:34:51.7508118Z * [new branch] piz/fall_back_missing_0716 -> origin/piz/fall_back_missing_0716 2025-07-17T08:34:51.7511349Z * [new branch] piz/fill_dist_cost_0702 -> origin/piz/fill_dist_cost_0702 2025-07-17T08:34:51.7514882Z * [new branch] piz/fill_dist_cost_0702-2 -> origin/piz/fill_dist_cost_0702-2 2025-07-17T08:34:51.7518257Z * [new branch] piz/fix_einsum -> origin/piz/fix_einsum 2025-07-17T08:34:51.7521548Z * [new branch] piz/fix_index_put_0706 -> origin/piz/fix_index_put_0706 2025-07-17T08:34:51.7524755Z * [new branch] piz/fix_slice -> origin/piz/fix_slice 2025-07-17T08:34:51.7528026Z * [new branch] piz/remove_assert_0709 -> origin/piz/remove_assert_0709 2025-07-17T08:34:51.7531442Z * [new branch] pool-separate -> origin/pool-separate 2025-07-17T08:34:51.7534715Z * [new branch] pr-156087 -> origin/pr-156087 2025-07-17T08:34:51.7539157Z * [new branch] pr/131860 -> origin/pr/131860 2025-07-17T08:34:51.7542383Z * [new branch] pr150241 -> origin/pr150241 2025-07-17T08:34:51.7545651Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-07-17T08:34:51.7549188Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-07-17T08:34:51.7554079Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-07-17T08:34:51.7557666Z * [new branch] randdcefix -> origin/randdcefix 2025-07-17T08:34:51.7562260Z * [new branch] release/1.10 -> origin/release/1.10 2025-07-17T08:34:51.7565647Z * [new branch] release/1.11 -> origin/release/1.11 2025-07-17T08:34:51.7569180Z * [new branch] release/1.12 -> origin/release/1.12 2025-07-17T08:34:51.7572374Z * [new branch] release/1.13 -> origin/release/1.13 2025-07-17T08:34:51.7575704Z * [new branch] release/1.4 -> origin/release/1.4 2025-07-17T08:34:51.7578754Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-07-17T08:34:51.7582170Z * [new branch] release/1.5 -> origin/release/1.5 2025-07-17T08:34:51.7585699Z * [new branch] release/1.6 -> origin/release/1.6 2025-07-17T08:34:51.7589147Z * [new branch] release/1.7 -> origin/release/1.7 2025-07-17T08:34:51.7592639Z * [new branch] release/1.8 -> origin/release/1.8 2025-07-17T08:34:51.7597576Z * [new branch] release/1.9 -> origin/release/1.9 2025-07-17T08:34:51.7600961Z * [new branch] release/2.0 -> origin/release/2.0 2025-07-17T08:34:51.7604463Z * [new branch] release/2.1 -> origin/release/2.1 2025-07-17T08:34:51.7607791Z * [new branch] release/2.2 -> origin/release/2.2 2025-07-17T08:34:51.7611270Z * [new branch] release/2.3 -> origin/release/2.3 2025-07-17T08:34:51.7614727Z * [new branch] release/2.4 -> origin/release/2.4 2025-07-17T08:34:51.7618034Z * [new branch] release/2.5 -> origin/release/2.5 2025-07-17T08:34:51.7621292Z * [new branch] release/2.6 -> origin/release/2.6 2025-07-17T08:34:51.7624642Z * [new branch] release/2.7 -> origin/release/2.7 2025-07-17T08:34:51.7628185Z * [new branch] release/2.8 -> origin/release/2.8 2025-07-17T08:34:51.7631467Z * [new branch] release_notes -> origin/release_notes 2025-07-17T08:34:51.7634871Z * [new branch] remove-actionable-label -> origin/remove-actionable-label 2025-07-17T08:34:51.7638393Z * [new branch] remove_ts_docs -> origin/remove_ts_docs 2025-07-17T08:34:51.7641995Z * [new branch] rephrase-export-remove-torchscript -> origin/rephrase-export-remove-torchscript 2025-07-17T08:34:51.7648501Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-07-17T08:34:51.7654903Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-07-17T08:34:51.7661284Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-07-17T08:34:51.7664803Z * [new branch] revert-156890-triton_34_update -> origin/revert-156890-triton_34_update 2025-07-17T08:34:51.7683852Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-07-17T08:34:51.7688170Z * [new branch] rithesh/fsdp_cpu -> origin/rithesh/fsdp_cpu 2025-07-17T08:34:51.7691636Z * [new branch] rocm-monitoring -> origin/rocm-monitoring 2025-07-17T08:34:51.7696275Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-07-17T08:34:51.7699431Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-07-17T08:34:51.7703877Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-07-17T08:34:51.7707189Z * [new branch] rzou/njt -> origin/rzou/njt 2025-07-17T08:34:51.7710428Z * [new branch] rzou/operator -> origin/rzou/operator 2025-07-17T08:34:51.7713806Z * [new branch] rzou/pca -> origin/rzou/pca 2025-07-17T08:34:51.7717397Z * [new branch] rzou/pipe_split -> origin/rzou/pipe_split 2025-07-17T08:34:51.7720397Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-07-17T08:34:51.7723728Z * [new branch] rzou/setup_context -> origin/rzou/setup_context 2025-07-17T08:34:51.7728411Z * [new branch] sanchitintel/refactor_aten_int8_woq_gemm -> origin/sanchitintel/refactor_aten_int8_woq_gemm 2025-07-17T08:34:51.7731784Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-07-17T08:34:51.7735123Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-07-17T08:34:51.7738472Z * [new branch] save -> origin/save 2025-07-17T08:34:51.7742978Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-07-17T08:34:51.7746510Z * [new branch] seemethere-patch-1 -> origin/seemethere-patch-1 2025-07-17T08:34:51.7749884Z * [new branch] share_and_pin_fork -> origin/share_and_pin_fork 2025-07-17T08:34:51.7754542Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-07-17T08:34:51.7758022Z * [new branch] shikaili_fp8_allgather -> origin/shikaili_fp8_allgather 2025-07-17T08:34:51.7761384Z * [new branch] shoumikhin-patch-12 -> origin/shoumikhin-patch-12 2025-07-17T08:34:51.7764827Z * [new branch] simplify-fq-per-channel -> origin/simplify-fq-per-channel 2025-07-17T08:34:51.7768209Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-07-17T08:34:51.7771669Z * [new branch] soumith-patch-1 -> origin/soumith-patch-1 2025-07-17T08:34:51.7776185Z * [new branch] sqzhang/flight4 -> origin/sqzhang/flight4 2025-07-17T08:34:51.7779455Z * [new branch] sqzhang/flight4plus -> origin/sqzhang/flight4plus 2025-07-17T08:34:51.7783908Z * [new branch] sraikund/record_funct_test -> origin/sraikund/record_funct_test 2025-07-17T08:34:51.7788336Z * [new branch] sraikund16/test -> origin/sraikund16/test 2025-07-17T08:34:51.7791786Z * [new branch] stablize-compilation-time -> origin/stablize-compilation-time 2025-07-17T08:34:51.7795328Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-07-17T08:34:51.7798697Z * [new branch] subgraph_fuse -> origin/subgraph_fuse 2025-07-17T08:34:51.7802168Z * [new branch] support-b200-runners -> origin/support-b200-runners 2025-07-17T08:34:51.7805610Z * [new branch] support-uv-in-collect_env -> origin/support-uv-in-collect_env 2025-07-17T08:34:51.7808942Z * [new branch] support_gqa_cpu -> origin/support_gqa_cpu 2025-07-17T08:34:51.7813488Z * [new branch] suryasub/fix-nccl-hang -> origin/suryasub/fix-nccl-hang 2025-07-17T08:34:51.7816730Z * [new branch] suryasub/nvshmem-get-kernel -> origin/suryasub/nvshmem-get-kernel 2025-07-17T08:34:51.7819931Z * [new branch] suryasub/put-signal-kernel -> origin/suryasub/put-signal-kernel 2025-07-17T08:34:51.7823266Z * [new branch] svekars-patch-1 -> origin/svekars-patch-1 2025-07-17T08:34:51.7826854Z * [new branch] svekars-patch-3 -> origin/svekars-patch-3 2025-07-17T08:34:51.7830257Z * [new branch] switch-bn -> origin/switch-bn 2025-07-17T08:34:51.7833661Z * [new branch] syed-ahmed-patch-1 -> origin/syed-ahmed-patch-1 2025-07-17T08:34:51.7841618Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-07-17T08:34:51.7846237Z * [new branch] teja/add_logs -> origin/teja/add_logs 2025-07-17T08:34:51.7849970Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-07-17T08:34:51.7853495Z * [new branch] test-internal-et -> origin/test-internal-et 2025-07-17T08:34:51.7856975Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-07-17T08:34:51.7861440Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-07-17T08:34:51.7864642Z * [new branch] test/inductor -> origin/test/inductor 2025-07-17T08:34:51.7868039Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-07-17T08:34:51.7871446Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-07-17T08:34:51.7875050Z * [new branch] tp_changes -> origin/tp_changes 2025-07-17T08:34:51.7878607Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-07-17T08:34:51.7881933Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-07-17T08:34:51.7885302Z * [new branch] trackMonitor -> origin/trackMonitor 2025-07-17T08:34:51.7888708Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-07-17T08:34:51.7892155Z * [new branch] tree_vec_base -> origin/tree_vec_base 2025-07-17T08:34:51.7895731Z * [new branch] triton-update -> origin/triton-update 2025-07-17T08:34:51.7899211Z * [new branch] triton_kernel -> origin/triton_kernel 2025-07-17T08:34:51.7902453Z * [new branch] triton_kernel_perf -> origin/triton_kernel_perf 2025-07-17T08:34:51.7905796Z * [new branch] try-runllm -> origin/try-runllm 2025-07-17T08:34:51.7909167Z * [new branch] type_dec -> origin/type_dec 2025-07-17T08:34:51.7912660Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-07-17T08:34:51.7917493Z * [new branch] update-audio-commit-hash/15573111403-1620-1 -> origin/update-audio-commit-hash/15573111403-1620-1 2025-07-17T08:34:51.7920699Z * [new branch] update-audio-commit-hash/15720911896-1627-1 -> origin/update-audio-commit-hash/15720911896-1627-1 2025-07-17T08:34:51.7923928Z * [new branch] update-audio-commit-hash/15890021967-1637-1 -> origin/update-audio-commit-hash/15890021967-1637-1 2025-07-17T08:34:51.7927213Z * [new branch] update-audio-commit-hash/16157311175-1652-1 -> origin/update-audio-commit-hash/16157311175-1652-1 2025-07-17T08:34:51.7930634Z * [new branch] update-audio-commit-hash/16307312222-1661-1 -> origin/update-audio-commit-hash/16307312222-1661-1 2025-07-17T08:34:51.7935233Z * [new branch] update-executorch-commit-hash/15694981040-1626-1 -> origin/update-executorch-commit-hash/15694981040-1626-1 2025-07-17T08:34:51.7939680Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-07-17T08:34:51.7944321Z * [new branch] update-vision-commit-hash/15336342773-1607-1 -> origin/update-vision-commit-hash/15336342773-1607-1 2025-07-17T08:34:51.7948761Z * [new branch] update-xla-commit-hash/15674789457-190-1 -> origin/update-xla-commit-hash/15674789457-190-1 2025-07-17T08:34:51.7952031Z * [new branch] update-xla-commit-hash/15818245712-191-1 -> origin/update-xla-commit-hash/15818245712-191-1 2025-07-17T08:34:51.7955426Z * [new branch] update-xla-commit-hash/16260974441-194-1 -> origin/update-xla-commit-hash/16260974441-194-1 2025-07-17T08:34:51.7959658Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-07-17T08:34:51.7963162Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-07-17T08:34:51.7966910Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-07-17T08:34:51.7970298Z * [new branch] update_slow_tests_1749454889 -> origin/update_slow_tests_1749454889 2025-07-17T08:34:51.7973675Z * [new branch] update_slow_tests_1751874113 -> origin/update_slow_tests_1751874113 2025-07-17T08:34:51.7977055Z * [new branch] update_slow_tests_1752478971 -> origin/update_slow_tests_1752478971 2025-07-17T08:34:51.7980504Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-07-17T08:34:51.7983757Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-07-17T08:34:51.7987254Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-07-17T08:34:51.7990845Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-07-17T08:34:51.7994558Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-07-17T08:34:51.7998073Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-07-17T08:34:51.8001547Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-07-17T08:34:51.8005606Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-07-17T08:34:51.8009071Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-07-17T08:34:51.8013087Z * [new branch] validate_fn -> origin/validate_fn 2025-07-17T08:34:51.8016636Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-07-17T08:34:51.8021111Z * [new branch] viable/strict -> origin/viable/strict 2025-07-17T08:34:51.8024473Z * [new branch] vkuzo-patch-1 -> origin/vkuzo-patch-1 2025-07-17T08:34:51.8028972Z * [new branch] wdvr/conda_devcontainer -> origin/wdvr/conda_devcontainer 2025-07-17T08:34:51.8032293Z * [new branch] wdvr/fix_logging_test -> origin/wdvr/fix_logging_test 2025-07-17T08:34:51.8035993Z * [new branch] wdvr/iss145259_alt -> origin/wdvr/iss145259_alt 2025-07-17T08:34:51.8039159Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-07-17T08:34:51.8042530Z * [new branch] weight_sharing_cpp -> origin/weight_sharing_cpp 2025-07-17T08:34:51.8047190Z * [new branch] whc/flight -> origin/whc/flight 2025-07-17T08:34:51.8050621Z * [new branch] whc/flight4 -> origin/whc/flight4 2025-07-17T08:34:51.8053856Z * [new branch] whc/flight51 -> origin/whc/flight51 2025-07-17T08:34:51.8057111Z * [new branch] whc/flight53 -> origin/whc/flight53 2025-07-17T08:34:51.8060474Z * [new branch] whc/flight_full -> origin/whc/flight_full 2025-07-17T08:34:51.8063662Z * [new branch] whc/p2phang -> origin/whc/p2phang 2025-07-17T08:34:51.8067054Z * [new branch] whc/stage2 -> origin/whc/stage2 2025-07-17T08:34:51.8070316Z * [new branch] whc/uneven -> origin/whc/uneven 2025-07-17T08:34:51.8073834Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-07-17T08:34:51.8077443Z * [new branch] win_warnings -> origin/win_warnings 2025-07-17T08:34:51.8081852Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-07-17T08:34:51.8085069Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-07-17T08:34:51.8088388Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-07-17T08:34:51.8091476Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-07-17T08:34:51.8095015Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-07-17T08:34:51.8098173Z * [new branch] xmfan/ca_api -> origin/xmfan/ca_api 2025-07-17T08:34:51.8101442Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-07-17T08:34:51.8104767Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-07-17T08:34:51.8108025Z * [new branch] xmfan/ca_cudagraphs -> origin/xmfan/ca_cudagraphs 2025-07-17T08:34:51.8111236Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-07-17T08:34:51.8114640Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-07-17T08:34:51.8117995Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-07-17T08:34:51.8121184Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-07-17T08:34:51.8124248Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-07-17T08:34:51.8127499Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-07-17T08:34:51.8130801Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-07-17T08:34:51.8134031Z * [new branch] xmfan/ca_mem_base -> origin/xmfan/ca_mem_base 2025-07-17T08:34:51.8137228Z * [new branch] xmfan/ca_mem_fix -> origin/xmfan/ca_mem_fix 2025-07-17T08:34:51.8140470Z * [new branch] xmfan/ca_memory_fix -> origin/xmfan/ca_memory_fix 2025-07-17T08:34:51.8143713Z * [new branch] xmfan/ca_memory_fix_rebased -> origin/xmfan/ca_memory_fix_rebased 2025-07-17T08:34:51.8147024Z * [new branch] xmfan/ca_memory_fix_rebased2 -> origin/xmfan/ca_memory_fix_rebased2 2025-07-17T08:34:51.8150276Z * [new branch] xmfan/ca_move_to_cuda -> origin/xmfan/ca_move_to_cuda 2025-07-17T08:34:51.8153643Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-07-17T08:34:51.8157200Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-07-17T08:34:51.8160410Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-07-17T08:34:51.8163593Z * [new branch] xmfan/ca_scalar -> origin/xmfan/ca_scalar 2025-07-17T08:34:51.8166972Z * [new branch] xmfan/ca_subclass_mem_fix -> origin/xmfan/ca_subclass_mem_fix 2025-07-17T08:34:51.8170164Z * [new branch] xmfan/ca_warm_mem -> origin/xmfan/ca_warm_mem 2025-07-17T08:34:51.8173406Z * [new branch] xmfan/ca_warm_mem_base -> origin/xmfan/ca_warm_mem_base 2025-07-17T08:34:51.8176546Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-07-17T08:34:51.8179821Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-07-17T08:34:51.8183175Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-07-17T08:34:51.8186368Z * [new branch] xmfan/cacu_may27 -> origin/xmfan/cacu_may27 2025-07-17T08:34:51.8189584Z * [new branch] xmfan/circular_dep -> origin/xmfan/circular_dep 2025-07-17T08:34:51.8192996Z * [new branch] xmfan/compiled_autograd_ddp -> origin/xmfan/compiled_autograd_ddp 2025-07-17T08:34:51.8196619Z * [new branch] xmfan/compiled_autograd_feb_29 -> origin/xmfan/compiled_autograd_feb_29 2025-07-17T08:34:51.8199997Z * [new branch] xmfan/compiled_autograd_graph_breaks -> origin/xmfan/compiled_autograd_graph_breaks 2025-07-17T08:34:51.8203555Z * [new branch] xmfan/compiled_autograd_hypothetical_perf -> origin/xmfan/compiled_autograd_hypothetical_perf 2025-07-17T08:34:51.8207011Z * [new branch] xmfan/compiled_autograd_perf_no_reuse -> origin/xmfan/compiled_autograd_perf_no_reuse 2025-07-17T08:34:51.8210218Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-07-17T08:34:51.8213467Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-07-17T08:34:51.8216641Z * [new branch] xmfan/feb_10_compiled_autograd -> origin/xmfan/feb_10_compiled_autograd 2025-07-17T08:34:51.8219956Z * [new branch] xmfan/feb_10_compiled_autograd_cudagraph -> origin/xmfan/feb_10_compiled_autograd_cudagraph 2025-07-17T08:34:51.8223096Z * [new branch] xmfan/issue_123374 -> origin/xmfan/issue_123374 2025-07-17T08:34:51.8227932Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-07-17T08:34:51.8231320Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-07-17T08:34:51.8234718Z * [new branch] xmfan/segfault_test -> origin/xmfan/segfault_test 2025-07-17T08:34:51.8238247Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-07-17T08:34:51.8242373Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-07-17T08:34:51.8244950Z * [new branch] xmfan/test -> origin/xmfan/test 2025-07-17T08:34:51.8248477Z * [new branch] y-do-we-have-7-build-systems -> origin/y-do-we-have-7-build-systems 2025-07-17T08:34:51.8253024Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-07-17T08:34:51.8256299Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-07-17T08:34:51.8259560Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-07-17T08:34:51.8262926Z * [new branch] yihan_quantization -> origin/yihan_quantization 2025-07-17T08:34:51.8267362Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-07-17T08:34:51.8271792Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-07-17T08:34:51.8275361Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-07-17T08:34:51.8278593Z * [new branch] zainr/mypy15-claude -> origin/zainr/mypy15-claude 2025-07-17T08:34:51.8281699Z * [new branch] zainr/pre-push-hooks -> origin/zainr/pre-push-hooks 2025-07-17T08:34:51.8285029Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-07-17T08:34:51.8288274Z * [new branch] zainr/test-workflow -> origin/zainr/test-workflow 2025-07-17T08:34:51.8291497Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-07-17T08:34:51.8294832Z * [new branch] zainr/unstable -> origin/zainr/unstable 2025-07-17T08:34:51.8298378Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-07-17T08:34:51.8301799Z * [new branch] zb2p -> origin/zb2p 2025-07-17T08:34:51.8305285Z * [new branch] zdevito-patch-1 -> origin/zdevito-patch-1 2025-07-17T08:34:51.8308757Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-07-17T08:34:51.8314334Z * [new branch] zhxchen17/nativert/0 -> origin/zhxchen17/nativert/0 2025-07-17T08:34:51.8318895Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-07-17T08:34:51.8323387Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-07-17T08:34:51.8327871Z * [new branch] zxiiro/bazel -> origin/zxiiro/bazel 2025-07-17T08:34:51.8331776Z * [new branch] zxiiro/docs -> origin/zxiiro/docs 2025-07-17T08:34:51.8334757Z * [new branch] zxiiro/linux-build -> origin/zxiiro/linux-build 2025-07-17T08:34:51.8338016Z * [new branch] zxiiro/linux-test -> origin/zxiiro/linux-test 2025-07-17T08:34:51.8341222Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-07-17T08:34:51.8344636Z * [new branch] zxiiro/test -> origin/zxiiro/test 2025-07-17T08:34:51.8347871Z * [new branch] zxiiro/windows -> origin/zxiiro/windows 2025-07-17T08:34:51.8351551Z * [new tag] bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug -> bc2caa7fdf006894eff7af936babde69ab5a40f8-huydhn-debug 2025-07-17T08:34:51.8354775Z * [new tag] ci/binaries/77164 -> ci/binaries/77164 2025-07-17T08:34:51.8357999Z * [new tag] ciflow/android/149601 -> ciflow/android/149601 2025-07-17T08:34:51.8360668Z * [new tag] ciflow/binaries/143959 -> ciflow/binaries/143959 2025-07-17T08:34:51.8362813Z * [new tag] ciflow/binaries/156049 -> ciflow/binaries/156049 2025-07-17T08:34:51.8364969Z * [new tag] ciflow/binaries/157432 -> ciflow/binaries/157432 2025-07-17T08:34:51.8367113Z * [new tag] ciflow/binaries/157685 -> ciflow/binaries/157685 2025-07-17T08:34:51.8369283Z * [new tag] ciflow/binaries/157689 -> ciflow/binaries/157689 2025-07-17T08:34:51.8371424Z * [new tag] ciflow/binaries/158104 -> ciflow/binaries/158104 2025-07-17T08:34:51.8373799Z * [new tag] ciflow/binaries/158151 -> ciflow/binaries/158151 2025-07-17T08:34:51.8376531Z * [new tag] ciflow/binaries_libtorch/143959 -> ciflow/binaries_libtorch/143959 2025-07-17T08:34:51.8378658Z * [new tag] ciflow/binaries_libtorch/156049 -> ciflow/binaries_libtorch/156049 2025-07-17T08:34:51.8380778Z * [new tag] ciflow/binaries_libtorch/157432 -> ciflow/binaries_libtorch/157432 2025-07-17T08:34:51.8383070Z * [new tag] ciflow/binaries_libtorch/157791 -> ciflow/binaries_libtorch/157791 2025-07-17T08:34:51.8385290Z * [new tag] ciflow/binaries_libtorch/157928 -> ciflow/binaries_libtorch/157928 2025-07-17T08:34:51.8387886Z * [new tag] ciflow/binaries_wheel/143959 -> ciflow/binaries_wheel/143959 2025-07-17T08:34:51.8390007Z * [new tag] ciflow/binaries_wheel/156049 -> ciflow/binaries_wheel/156049 2025-07-17T08:34:51.8392155Z * [new tag] ciflow/binaries_wheel/157432 -> ciflow/binaries_wheel/157432 2025-07-17T08:34:51.8394366Z * [new tag] ciflow/binaries_wheel/157791 -> ciflow/binaries_wheel/157791 2025-07-17T08:34:51.8396585Z * [new tag] ciflow/binaries_wheel/157928 -> ciflow/binaries_wheel/157928 2025-07-17T08:34:51.8398807Z * [new tag] ciflow/binaries_wheel/158408 -> ciflow/binaries_wheel/158408 2025-07-17T08:34:51.8401592Z * [new tag] ciflow/h100-cutlass-backend/156626 -> ciflow/h100-cutlass-backend/156626 2025-07-17T08:34:51.8404165Z * [new tag] ciflow/h100-distributed/144552 -> ciflow/h100-distributed/144552 2025-07-17T08:34:51.8406273Z * [new tag] ciflow/h100-distributed/150312 -> ciflow/h100-distributed/150312 2025-07-17T08:34:51.8408388Z * [new tag] ciflow/h100-distributed/156605 -> ciflow/h100-distributed/156605 2025-07-17T08:34:51.8410723Z * [new tag] ciflow/h100-distributed/156703 -> ciflow/h100-distributed/156703 2025-07-17T08:34:51.8413418Z * [new tag] ciflow/h100-symm-mem/151845 -> ciflow/h100-symm-mem/151845 2025-07-17T08:34:51.8415556Z * [new tag] ciflow/h100-symm-mem/157180 -> ciflow/h100-symm-mem/157180 2025-07-17T08:34:51.8417662Z * [new tag] ciflow/h100-symm-mem/157970 -> ciflow/h100-symm-mem/157970 2025-07-17T08:34:51.8420204Z * [new tag] ciflow/h100-symm-mem/158511 -> ciflow/h100-symm-mem/158511 2025-07-17T08:34:51.8422097Z * [new tag] ciflow/h100-symm-mem/158512 -> ciflow/h100-symm-mem/158512 2025-07-17T08:34:51.8424245Z * [new tag] ciflow/h100-symm-mem/158513 -> ciflow/h100-symm-mem/158513 2025-07-17T08:34:51.8426480Z * [new tag] ciflow/h100-symm-mem/158514 -> ciflow/h100-symm-mem/158514 2025-07-17T08:34:51.8428630Z * [new tag] ciflow/h100-symm-mem/158515 -> ciflow/h100-symm-mem/158515 2025-07-17T08:34:51.8430991Z * [new tag] ciflow/h100-symm-mem/158523 -> ciflow/h100-symm-mem/158523 2025-07-17T08:34:51.8433592Z * [new tag] ciflow/h100/156980 -> ciflow/h100/156980 2025-07-17T08:34:51.8438290Z * [new tag] ciflow/h100/158459 -> ciflow/h100/158459 2025-07-17T08:34:51.8441090Z * [new tag] ciflow/inductor-perf-test-nightly-rocm/151845 -> ciflow/inductor-perf-test-nightly-rocm/151845 2025-07-17T08:34:51.8443391Z * [new tag] ciflow/inductor-perf-test-nightly-rocm/156592 -> ciflow/inductor-perf-test-nightly-rocm/156592 2025-07-17T08:34:51.8446020Z * [new tag] ciflow/inductor-perf-test-nightly/156592 -> ciflow/inductor-perf-test-nightly/156592 2025-07-17T08:34:51.8448514Z * [new tag] ciflow/inductor-rocm/151845 -> ciflow/inductor-rocm/151845 2025-07-17T08:34:51.8450877Z * [new tag] ciflow/inductor-rocm/156192 -> ciflow/inductor-rocm/156192 2025-07-17T08:34:51.8453156Z * [new tag] ciflow/inductor-rocm/157191 -> ciflow/inductor-rocm/157191 2025-07-17T08:34:51.8455291Z * [new tag] ciflow/inductor-rocm/157520 -> ciflow/inductor-rocm/157520 2025-07-17T08:34:51.8457436Z * [new tag] ciflow/inductor-rocm/158074 -> ciflow/inductor-rocm/158074 2025-07-17T08:34:51.8459691Z * [new tag] ciflow/inductor-rocm/158102 -> ciflow/inductor-rocm/158102 2025-07-17T08:34:51.8461974Z * [new tag] ciflow/inductor-rocm/158103 -> ciflow/inductor-rocm/158103 2025-07-17T08:34:51.8464232Z * [new tag] ciflow/inductor-rocm/158459 -> ciflow/inductor-rocm/158459 2025-07-17T08:34:51.8466833Z * [new tag] ciflow/inductor/137400 -> ciflow/inductor/137400 2025-07-17T08:34:51.8468956Z * [new tag] ciflow/inductor/138214 -> ciflow/inductor/138214 2025-07-17T08:34:51.8471111Z * [new tag] ciflow/inductor/139561 -> ciflow/inductor/139561 2025-07-17T08:34:51.8473230Z * [new tag] ciflow/inductor/143712 -> ciflow/inductor/143712 2025-07-17T08:34:51.8475611Z * [new tag] ciflow/inductor/144516 -> ciflow/inductor/144516 2025-07-17T08:34:51.8477832Z * [new tag] ciflow/inductor/144556 -> ciflow/inductor/144556 2025-07-17T08:34:51.8479945Z * [new tag] ciflow/inductor/147470 -> ciflow/inductor/147470 2025-07-17T08:34:51.8482122Z * [new tag] ciflow/inductor/148180 -> ciflow/inductor/148180 2025-07-17T08:34:51.8484316Z * [new tag] ciflow/inductor/148328 -> ciflow/inductor/148328 2025-07-17T08:34:51.8486486Z * [new tag] ciflow/inductor/148484 -> ciflow/inductor/148484 2025-07-17T08:34:51.8488636Z * [new tag] ciflow/inductor/148492 -> ciflow/inductor/148492 2025-07-17T08:34:51.8490740Z * [new tag] ciflow/inductor/148569 -> ciflow/inductor/148569 2025-07-17T08:34:51.8492897Z * [new tag] ciflow/inductor/149003 -> ciflow/inductor/149003 2025-07-17T08:34:51.8495014Z * [new tag] ciflow/inductor/149961 -> ciflow/inductor/149961 2025-07-17T08:34:51.8497211Z * [new tag] ciflow/inductor/150762 -> ciflow/inductor/150762 2025-07-17T08:34:51.8499473Z * [new tag] ciflow/inductor/151777 -> ciflow/inductor/151777 2025-07-17T08:34:51.8501815Z * [new tag] ciflow/inductor/151845 -> ciflow/inductor/151845 2025-07-17T08:34:51.8503738Z * [new tag] ciflow/inductor/152624 -> ciflow/inductor/152624 2025-07-17T08:34:51.8505856Z * [new tag] ciflow/inductor/154149 -> ciflow/inductor/154149 2025-07-17T08:34:51.8507988Z * [new tag] ciflow/inductor/154199 -> ciflow/inductor/154199 2025-07-17T08:34:51.8510122Z * [new tag] ciflow/inductor/154551 -> ciflow/inductor/154551 2025-07-17T08:34:51.8512234Z * [new tag] ciflow/inductor/154694 -> ciflow/inductor/154694 2025-07-17T08:34:51.8514551Z * [new tag] ciflow/inductor/155452 -> ciflow/inductor/155452 2025-07-17T08:34:51.8516752Z * [new tag] ciflow/inductor/155608 -> ciflow/inductor/155608 2025-07-17T08:34:51.8518882Z * [new tag] ciflow/inductor/155877 -> ciflow/inductor/155877 2025-07-17T08:34:51.8521133Z * [new tag] ciflow/inductor/155958 -> ciflow/inductor/155958 2025-07-17T08:34:51.8523277Z * [new tag] ciflow/inductor/156049 -> ciflow/inductor/156049 2025-07-17T08:34:51.8525398Z * [new tag] ciflow/inductor/156118 -> ciflow/inductor/156118 2025-07-17T08:34:51.8527572Z * [new tag] ciflow/inductor/156141 -> ciflow/inductor/156141 2025-07-17T08:34:51.8529707Z * [new tag] ciflow/inductor/156175 -> ciflow/inductor/156175 2025-07-17T08:34:51.8532013Z * [new tag] ciflow/inductor/156192 -> ciflow/inductor/156192 2025-07-17T08:34:51.8534255Z * [new tag] ciflow/inductor/156296 -> ciflow/inductor/156296 2025-07-17T08:34:51.8536539Z * [new tag] ciflow/inductor/156369 -> ciflow/inductor/156369 2025-07-17T08:34:51.8538538Z * [new tag] ciflow/inductor/156370 -> ciflow/inductor/156370 2025-07-17T08:34:51.8540680Z * [new tag] ciflow/inductor/156371 -> ciflow/inductor/156371 2025-07-17T08:34:51.8542811Z * [new tag] ciflow/inductor/156416 -> ciflow/inductor/156416 2025-07-17T08:34:51.8544941Z * [new tag] ciflow/inductor/156460 -> ciflow/inductor/156460 2025-07-17T08:34:51.8547086Z * [new tag] ciflow/inductor/156592 -> ciflow/inductor/156592 2025-07-17T08:34:51.8549276Z * [new tag] ciflow/inductor/156605 -> ciflow/inductor/156605 2025-07-17T08:34:51.8551414Z * [new tag] ciflow/inductor/156626 -> ciflow/inductor/156626 2025-07-17T08:34:51.8553516Z * [new tag] ciflow/inductor/156652 -> ciflow/inductor/156652 2025-07-17T08:34:51.8555797Z * [new tag] ciflow/inductor/156781 -> ciflow/inductor/156781 2025-07-17T08:34:51.8557961Z * [new tag] ciflow/inductor/156851 -> ciflow/inductor/156851 2025-07-17T08:34:51.8560347Z * [new tag] ciflow/inductor/156874 -> ciflow/inductor/156874 2025-07-17T08:34:51.8562510Z * [new tag] ciflow/inductor/156977 -> ciflow/inductor/156977 2025-07-17T08:34:51.8564653Z * [new tag] ciflow/inductor/156980 -> ciflow/inductor/156980 2025-07-17T08:34:51.8566906Z * [new tag] ciflow/inductor/157152 -> ciflow/inductor/157152 2025-07-17T08:34:51.8569097Z * [new tag] ciflow/inductor/157298 -> ciflow/inductor/157298 2025-07-17T08:34:51.8571197Z * [new tag] ciflow/inductor/157520 -> ciflow/inductor/157520 2025-07-17T08:34:51.8573353Z * [new tag] ciflow/inductor/157572 -> ciflow/inductor/157572 2025-07-17T08:34:51.8575514Z * [new tag] ciflow/inductor/157580 -> ciflow/inductor/157580 2025-07-17T08:34:51.8577647Z * [new tag] ciflow/inductor/157594 -> ciflow/inductor/157594 2025-07-17T08:34:51.8580173Z * [new tag] ciflow/inductor/157633 -> ciflow/inductor/157633 2025-07-17T08:34:51.8582150Z * [new tag] ciflow/inductor/157635 -> ciflow/inductor/157635 2025-07-17T08:34:51.8584319Z * [new tag] ciflow/inductor/157685 -> ciflow/inductor/157685 2025-07-17T08:34:51.8586632Z * [new tag] ciflow/inductor/157686 -> ciflow/inductor/157686 2025-07-17T08:34:51.8588770Z * [new tag] ciflow/inductor/157689 -> ciflow/inductor/157689 2025-07-17T08:34:51.8590889Z * [new tag] ciflow/inductor/157699 -> ciflow/inductor/157699 2025-07-17T08:34:51.8593161Z * [new tag] ciflow/inductor/157743 -> ciflow/inductor/157743 2025-07-17T08:34:51.8595575Z * [new tag] ciflow/inductor/157748 -> ciflow/inductor/157748 2025-07-17T08:34:51.8597819Z * [new tag] ciflow/inductor/157804 -> ciflow/inductor/157804 2025-07-17T08:34:51.8599894Z * [new tag] ciflow/inductor/157822 -> ciflow/inductor/157822 2025-07-17T08:34:51.8602364Z * [new tag] ciflow/inductor/157854 -> ciflow/inductor/157854 2025-07-17T08:34:51.8604781Z * [new tag] ciflow/inductor/157887 -> ciflow/inductor/157887 2025-07-17T08:34:51.8606931Z * [new tag] ciflow/inductor/157902 -> ciflow/inductor/157902 2025-07-17T08:34:51.8609088Z * [new tag] ciflow/inductor/157927 -> ciflow/inductor/157927 2025-07-17T08:34:51.8611253Z * [new tag] ciflow/inductor/157944 -> ciflow/inductor/157944 2025-07-17T08:34:51.8613553Z * [new tag] ciflow/inductor/157951 -> ciflow/inductor/157951 2025-07-17T08:34:51.8615948Z * [new tag] ciflow/inductor/157954 -> ciflow/inductor/157954 2025-07-17T08:34:51.8618342Z * [new tag] ciflow/inductor/157967 -> ciflow/inductor/157967 2025-07-17T08:34:51.8620558Z * [new tag] ciflow/inductor/157969 -> ciflow/inductor/157969 2025-07-17T08:34:51.8622696Z * [new tag] ciflow/inductor/157971 -> ciflow/inductor/157971 2025-07-17T08:34:51.8624830Z * [new tag] ciflow/inductor/157979 -> ciflow/inductor/157979 2025-07-17T08:34:51.8627127Z * [new tag] ciflow/inductor/157982 -> ciflow/inductor/157982 2025-07-17T08:34:51.8629283Z * [new tag] ciflow/inductor/157993 -> ciflow/inductor/157993 2025-07-17T08:34:51.8631424Z * [new tag] ciflow/inductor/158004 -> ciflow/inductor/158004 2025-07-17T08:34:51.8633562Z * [new tag] ciflow/inductor/158011 -> ciflow/inductor/158011 2025-07-17T08:34:51.8635872Z * [new tag] ciflow/inductor/158015 -> ciflow/inductor/158015 2025-07-17T08:34:51.8638134Z * [new tag] ciflow/inductor/158046 -> ciflow/inductor/158046 2025-07-17T08:34:51.8640387Z * [new tag] ciflow/inductor/158047 -> ciflow/inductor/158047 2025-07-17T08:34:51.8642494Z * [new tag] ciflow/inductor/158048 -> ciflow/inductor/158048 2025-07-17T08:34:51.8644632Z * [new tag] ciflow/inductor/158049 -> ciflow/inductor/158049 2025-07-17T08:34:51.8646743Z * [new tag] ciflow/inductor/158061 -> ciflow/inductor/158061 2025-07-17T08:34:51.8648909Z * [new tag] ciflow/inductor/158062 -> ciflow/inductor/158062 2025-07-17T08:34:51.8651034Z * [new tag] ciflow/inductor/158072 -> ciflow/inductor/158072 2025-07-17T08:34:51.8653155Z * [new tag] ciflow/inductor/158074 -> ciflow/inductor/158074 2025-07-17T08:34:51.8655373Z * [new tag] ciflow/inductor/158075 -> ciflow/inductor/158075 2025-07-17T08:34:51.8657524Z * [new tag] ciflow/inductor/158084 -> ciflow/inductor/158084 2025-07-17T08:34:51.8659631Z * [new tag] ciflow/inductor/158091 -> ciflow/inductor/158091 2025-07-17T08:34:51.8661917Z * [new tag] ciflow/inductor/158097 -> ciflow/inductor/158097 2025-07-17T08:34:51.8663928Z * [new tag] ciflow/inductor/158098 -> ciflow/inductor/158098 2025-07-17T08:34:51.8666106Z * [new tag] ciflow/inductor/158104 -> ciflow/inductor/158104 2025-07-17T08:34:51.8668193Z * [new tag] ciflow/inductor/158119 -> ciflow/inductor/158119 2025-07-17T08:34:51.8670417Z * [new tag] ciflow/inductor/158156 -> ciflow/inductor/158156 2025-07-17T08:34:51.8672735Z * [new tag] ciflow/inductor/158171 -> ciflow/inductor/158171 2025-07-17T08:34:51.8676213Z * [new tag] ciflow/inductor/158174 -> ciflow/inductor/158174 2025-07-17T08:34:51.8678341Z * [new tag] ciflow/inductor/158188 -> ciflow/inductor/158188 2025-07-17T08:34:51.8680486Z * [new tag] ciflow/inductor/158193 -> ciflow/inductor/158193 2025-07-17T08:34:51.8682657Z * [new tag] ciflow/inductor/158211 -> ciflow/inductor/158211 2025-07-17T08:34:51.8684766Z * [new tag] ciflow/inductor/158223 -> ciflow/inductor/158223 2025-07-17T08:34:51.8686928Z * [new tag] ciflow/inductor/158237 -> ciflow/inductor/158237 2025-07-17T08:34:51.8689054Z * [new tag] ciflow/inductor/158259 -> ciflow/inductor/158259 2025-07-17T08:34:51.8691404Z * [new tag] ciflow/inductor/158284 -> ciflow/inductor/158284 2025-07-17T08:34:51.8693649Z * [new tag] ciflow/inductor/158289 -> ciflow/inductor/158289 2025-07-17T08:34:51.8695792Z * [new tag] ciflow/inductor/158311 -> ciflow/inductor/158311 2025-07-17T08:34:51.8698058Z * [new tag] ciflow/inductor/158312 -> ciflow/inductor/158312 2025-07-17T08:34:51.8700222Z * [new tag] ciflow/inductor/158318 -> ciflow/inductor/158318 2025-07-17T08:34:51.8702382Z * [new tag] ciflow/inductor/158321 -> ciflow/inductor/158321 2025-07-17T08:34:51.8704632Z * [new tag] ciflow/inductor/158338 -> ciflow/inductor/158338 2025-07-17T08:34:51.8706811Z * [new tag] ciflow/inductor/158349 -> ciflow/inductor/158349 2025-07-17T08:34:51.8708987Z * [new tag] ciflow/inductor/158350 -> ciflow/inductor/158350 2025-07-17T08:34:51.8711124Z * [new tag] ciflow/inductor/158351 -> ciflow/inductor/158351 2025-07-17T08:34:51.8713276Z * [new tag] ciflow/inductor/158360 -> ciflow/inductor/158360 2025-07-17T08:34:51.8715546Z * [new tag] ciflow/inductor/158361 -> ciflow/inductor/158361 2025-07-17T08:34:51.8717738Z * [new tag] ciflow/inductor/158363 -> ciflow/inductor/158363 2025-07-17T08:34:51.8719853Z * [new tag] ciflow/inductor/158368 -> ciflow/inductor/158368 2025-07-17T08:34:51.8721999Z * [new tag] ciflow/inductor/158377 -> ciflow/inductor/158377 2025-07-17T08:34:51.8724156Z * [new tag] ciflow/inductor/158379 -> ciflow/inductor/158379 2025-07-17T08:34:51.8726383Z * [new tag] ciflow/inductor/158381 -> ciflow/inductor/158381 2025-07-17T08:34:51.8728529Z * [new tag] ciflow/inductor/158391 -> ciflow/inductor/158391 2025-07-17T08:34:51.8730861Z * [new tag] ciflow/inductor/158394 -> ciflow/inductor/158394 2025-07-17T08:34:51.8733074Z * [new tag] ciflow/inductor/158397 -> ciflow/inductor/158397 2025-07-17T08:34:51.8735394Z * [new tag] ciflow/inductor/158400 -> ciflow/inductor/158400 2025-07-17T08:34:51.8737677Z * [new tag] ciflow/inductor/158404 -> ciflow/inductor/158404 2025-07-17T08:34:51.8739873Z * [new tag] ciflow/inductor/158406 -> ciflow/inductor/158406 2025-07-17T08:34:51.8742186Z * [new tag] ciflow/inductor/158410 -> ciflow/inductor/158410 2025-07-17T08:34:51.8744347Z * [new tag] ciflow/inductor/158418 -> ciflow/inductor/158418 2025-07-17T08:34:51.8746582Z * [new tag] ciflow/inductor/158424 -> ciflow/inductor/158424 2025-07-17T08:34:51.8748743Z * [new tag] ciflow/inductor/158426 -> ciflow/inductor/158426 2025-07-17T08:34:51.8750879Z * [new tag] ciflow/inductor/158427 -> ciflow/inductor/158427 2025-07-17T08:34:51.8753301Z * [new tag] ciflow/inductor/158430 -> ciflow/inductor/158430 2025-07-17T08:34:51.8755764Z * [new tag] ciflow/inductor/158435 -> ciflow/inductor/158435 2025-07-17T08:34:51.8757952Z * [new tag] ciflow/inductor/158442 -> ciflow/inductor/158442 2025-07-17T08:34:51.8760159Z * [new tag] ciflow/inductor/158449 -> ciflow/inductor/158449 2025-07-17T08:34:51.8762407Z * [new tag] ciflow/inductor/158456 -> ciflow/inductor/158456 2025-07-17T08:34:51.8764591Z * [new tag] ciflow/inductor/158458 -> ciflow/inductor/158458 2025-07-17T08:34:51.8766715Z * [new tag] ciflow/inductor/158459 -> ciflow/inductor/158459 2025-07-17T08:34:51.8768854Z * [new tag] ciflow/inductor/158460 -> ciflow/inductor/158460 2025-07-17T08:34:51.8771045Z * [new tag] ciflow/inductor/158462 -> ciflow/inductor/158462 2025-07-17T08:34:51.8773127Z * [new tag] ciflow/inductor/158467 -> ciflow/inductor/158467 2025-07-17T08:34:51.8775312Z * [new tag] ciflow/inductor/158468 -> ciflow/inductor/158468 2025-07-17T08:34:51.8777432Z * [new tag] ciflow/inductor/158476 -> ciflow/inductor/158476 2025-07-17T08:34:51.8779885Z * [new tag] ciflow/inductor/158479 -> ciflow/inductor/158479 2025-07-17T08:34:51.8782050Z * [new tag] ciflow/inductor/158480 -> ciflow/inductor/158480 2025-07-17T08:34:51.8784203Z * [new tag] ciflow/inductor/158485 -> ciflow/inductor/158485 2025-07-17T08:34:51.8786354Z * [new tag] ciflow/inductor/158490 -> ciflow/inductor/158490 2025-07-17T08:34:51.8788498Z * [new tag] ciflow/inductor/158492 -> ciflow/inductor/158492 2025-07-17T08:34:51.8790829Z * [new tag] ciflow/inductor/158494 -> ciflow/inductor/158494 2025-07-17T08:34:51.8792970Z * [new tag] ciflow/inductor/158495 -> ciflow/inductor/158495 2025-07-17T08:34:51.8795348Z * [new tag] ciflow/inductor/158498 -> ciflow/inductor/158498 2025-07-17T08:34:51.8797570Z * [new tag] ciflow/inductor/158499 -> ciflow/inductor/158499 2025-07-17T08:34:51.8799707Z * [new tag] ciflow/inductor/158500 -> ciflow/inductor/158500 2025-07-17T08:34:51.8801808Z * [new tag] ciflow/inductor/158501 -> ciflow/inductor/158501 2025-07-17T08:34:51.8803968Z * [new tag] ciflow/inductor/158504 -> ciflow/inductor/158504 2025-07-17T08:34:51.8806258Z * [new tag] ciflow/inductor/158509 -> ciflow/inductor/158509 2025-07-17T08:34:51.8808231Z * [new tag] ciflow/inductor/158520 -> ciflow/inductor/158520 2025-07-17T08:34:51.8810350Z * [new tag] ciflow/inductor/158524 -> ciflow/inductor/158524 2025-07-17T08:34:51.8812606Z * [new tag] ciflow/inductor/158525 -> ciflow/inductor/158525 2025-07-17T08:34:51.8814920Z * [new tag] ciflow/inductor/158526 -> ciflow/inductor/158526 2025-07-17T08:34:51.8817070Z * [new tag] ciflow/inductor/158528 -> ciflow/inductor/158528 2025-07-17T08:34:51.8819201Z * [new tag] ciflow/inductor/158534 -> ciflow/inductor/158534 2025-07-17T08:34:51.8821459Z * [new tag] ciflow/inductor/158535 -> ciflow/inductor/158535 2025-07-17T08:34:51.8823924Z * [new tag] ciflow/inductor/158537 -> ciflow/inductor/158537 2025-07-17T08:34:51.8825768Z * [new tag] ciflow/inductor/158538 -> ciflow/inductor/158538 2025-07-17T08:34:51.8828076Z * [new tag] ciflow/inductor/158543 -> ciflow/inductor/158543 2025-07-17T08:34:51.8830549Z * [new tag] ciflow/inductor/3b9a386 -> ciflow/inductor/3b9a386 2025-07-17T08:34:51.8833010Z * [new tag] ciflow/inductor/3d4b92b -> ciflow/inductor/3d4b92b 2025-07-17T08:34:51.8835430Z * [new tag] ciflow/inductor/d224ac7 -> ciflow/inductor/d224ac7 2025-07-17T08:34:51.8838087Z * [new tag] ciflow/linux-aarch64/157520 -> ciflow/linux-aarch64/157520 2025-07-17T08:34:51.8840178Z * [new tag] ciflow/linux-aarch64/157782 -> ciflow/linux-aarch64/157782 2025-07-17T08:34:51.8842417Z * [new tag] ciflow/linux-aarch64/157994 -> ciflow/linux-aarch64/157994 2025-07-17T08:34:51.8844721Z * [new tag] ciflow/linux-aarch64/158445 -> ciflow/linux-aarch64/158445 2025-07-17T08:34:51.8847300Z * [new tag] ciflow/mps/155200 -> ciflow/mps/155200 2025-07-17T08:34:51.8849465Z * [new tag] ciflow/mps/157553 -> ciflow/mps/157553 2025-07-17T08:34:51.8851601Z * [new tag] ciflow/mps/157875 -> ciflow/mps/157875 2025-07-17T08:34:51.8853796Z * [new tag] ciflow/mps/157876 -> ciflow/mps/157876 2025-07-17T08:34:51.8855939Z * [new tag] ciflow/mps/158237 -> ciflow/mps/158237 2025-07-17T08:34:51.8858065Z * [new tag] ciflow/mps/158349 -> ciflow/mps/158349 2025-07-17T08:34:51.8860251Z * [new tag] ciflow/mps/158350 -> ciflow/mps/158350 2025-07-17T08:34:51.8862387Z * [new tag] ciflow/mps/158351 -> ciflow/mps/158351 2025-07-17T08:34:51.8865064Z * [new tag] ciflow/periodic-rocm-mi300/156192 -> ciflow/periodic-rocm-mi300/156192 2025-07-17T08:34:51.8867177Z * [new tag] ciflow/periodic-rocm-mi300/157191 -> ciflow/periodic-rocm-mi300/157191 2025-07-17T08:34:51.8869291Z * [new tag] ciflow/periodic-rocm-mi300/158102 -> ciflow/periodic-rocm-mi300/158102 2025-07-17T08:34:51.8871414Z * [new tag] ciflow/periodic-rocm-mi300/158103 -> ciflow/periodic-rocm-mi300/158103 2025-07-17T08:34:51.8874377Z * [new tag] ciflow/periodic/054a2fd -> ciflow/periodic/054a2fd 2025-07-17T08:34:51.8876645Z * [new tag] ciflow/periodic/143959 -> ciflow/periodic/143959 2025-07-17T08:34:51.8878973Z * [new tag] ciflow/periodic/156559 -> ciflow/periodic/156559 2025-07-17T08:34:51.8881234Z * [new tag] ciflow/periodic/156900 -> ciflow/periodic/156900 2025-07-17T08:34:51.8883362Z * [new tag] ciflow/periodic/157748 -> ciflow/periodic/157748 2025-07-17T08:34:51.8885641Z * [new tag] ciflow/periodic/157939 -> ciflow/periodic/157939 2025-07-17T08:34:51.8887901Z * [new tag] ciflow/periodic/158145 -> ciflow/periodic/158145 2025-07-17T08:34:51.8890219Z * [new tag] ciflow/periodic/2a6d37d -> ciflow/periodic/2a6d37d 2025-07-17T08:34:51.8892512Z * [new tag] ciflow/periodic/317eeb8 -> ciflow/periodic/317eeb8 2025-07-17T08:34:51.8894801Z * [new tag] ciflow/periodic/3c32 -> ciflow/periodic/3c32 2025-07-17T08:34:51.8897174Z * [new tag] ciflow/periodic/3e98831 -> ciflow/periodic/3e98831 2025-07-17T08:34:51.8899514Z * [new tag] ciflow/periodic/94512-point -> ciflow/periodic/94512-point 2025-07-17T08:34:51.8902436Z * [new tag] ciflow/periodic/csl/test87519 -> ciflow/periodic/csl/test87519 2025-07-17T08:34:51.8904758Z * [new tag] ciflow/periodic/csltest88275 -> ciflow/periodic/csltest88275 2025-07-17T08:34:51.8907229Z * [new tag] ciflow/periodic/csltest88761 -> ciflow/periodic/csltest88761 2025-07-17T08:34:51.8909419Z * [new tag] ciflow/periodic/release_1.12 -> ciflow/periodic/release_1.12 2025-07-17T08:34:51.8911894Z * [new tag] ciflow/periodic/release_1.12.0 -> ciflow/periodic/release_1.12.0 2025-07-17T08:34:51.8915777Z * [new tag] ciflow/periodic/sha-ec5b83 -> ciflow/periodic/sha-ec5b83 2025-07-17T08:34:51.8919487Z * [new tag] ciflow/rocm-mi300/156192 -> ciflow/rocm-mi300/156192 2025-07-17T08:34:51.8920865Z * [new tag] ciflow/rocm-mi300/157191 -> ciflow/rocm-mi300/157191 2025-07-17T08:34:51.8922836Z * [new tag] ciflow/rocm-mi300/157520 -> ciflow/rocm-mi300/157520 2025-07-17T08:34:51.8924984Z * [new tag] ciflow/rocm-mi300/158102 -> ciflow/rocm-mi300/158102 2025-07-17T08:34:51.8927091Z * [new tag] ciflow/rocm-mi300/158103 -> ciflow/rocm-mi300/158103 2025-07-17T08:34:51.8929324Z * [new tag] ciflow/rocm-mi300/158221 -> ciflow/rocm-mi300/158221 2025-07-17T08:34:51.8931500Z * [new tag] ciflow/rocm-mi300/158459 -> ciflow/rocm-mi300/158459 2025-07-17T08:34:51.8934106Z * [new tag] ciflow/rocm/148492 -> ciflow/rocm/148492 2025-07-17T08:34:51.8936216Z * [new tag] ciflow/rocm/149601 -> ciflow/rocm/149601 2025-07-17T08:34:51.8938386Z * [new tag] ciflow/rocm/150312 -> ciflow/rocm/150312 2025-07-17T08:34:51.8940474Z * [new tag] ciflow/rocm/151845 -> ciflow/rocm/151845 2025-07-17T08:34:51.8942656Z * [new tag] ciflow/rocm/155200 -> ciflow/rocm/155200 2025-07-17T08:34:51.8944785Z * [new tag] ciflow/rocm/155877 -> ciflow/rocm/155877 2025-07-17T08:34:51.8947072Z * [new tag] ciflow/rocm/156165 -> ciflow/rocm/156165 2025-07-17T08:34:51.8949206Z * [new tag] ciflow/rocm/156192 -> ciflow/rocm/156192 2025-07-17T08:34:51.8951339Z * [new tag] ciflow/rocm/156592 -> ciflow/rocm/156592 2025-07-17T08:34:51.8953442Z * [new tag] ciflow/rocm/157520 -> ciflow/rocm/157520 2025-07-17T08:34:51.8956101Z * [new tag] ciflow/rocm/157964 -> ciflow/rocm/157964 2025-07-17T08:34:51.8958302Z * [new tag] ciflow/rocm/158074 -> ciflow/rocm/158074 2025-07-17T08:34:51.8960448Z * [new tag] ciflow/rocm/158219 -> ciflow/rocm/158219 2025-07-17T08:34:51.8962565Z * [new tag] ciflow/rocm/158220 -> ciflow/rocm/158220 2025-07-17T08:34:51.8964777Z * [new tag] ciflow/rocm/158224 -> ciflow/rocm/158224 2025-07-17T08:34:51.8967033Z * [new tag] ciflow/rocm/158271 -> ciflow/rocm/158271 2025-07-17T08:34:51.8969219Z * [new tag] ciflow/rocm/158408 -> ciflow/rocm/158408 2025-07-17T08:34:51.8971356Z * [new tag] ciflow/rocm/158459 -> ciflow/rocm/158459 2025-07-17T08:34:51.8973956Z * [new tag] ciflow/s390/143959 -> ciflow/s390/143959 2025-07-17T08:34:51.8976225Z * [new tag] ciflow/s390/151447 -> ciflow/s390/151447 2025-07-17T08:34:51.8978977Z * [new tag] ciflow/slow/01c7106 -> ciflow/slow/01c7106 2025-07-17T08:34:51.8981233Z * [new tag] ciflow/slow/0577043 -> ciflow/slow/0577043 2025-07-17T08:34:51.8984411Z * [new tag] ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym -> ciflow/slow/0d5b74da0cab798fbfdb9caa53fad816999c8386-sdym 2025-07-17T08:34:51.8986626Z * [new tag] ciflow/slow/0e81104 -> ciflow/slow/0e81104 2025-07-17T08:34:51.8989073Z * [new tag] ciflow/slow/157385 -> ciflow/slow/157385 2025-07-17T08:34:51.8991460Z * [new tag] ciflow/slow/157748 -> ciflow/slow/157748 2025-07-17T08:34:51.8993552Z * [new tag] ciflow/slow/158222 -> ciflow/slow/158222 2025-07-17T08:34:51.8995997Z * [new tag] ciflow/slow/158312 -> ciflow/slow/158312 2025-07-17T08:34:51.8998273Z * [new tag] ciflow/slow/158424 -> ciflow/slow/158424 2025-07-17T08:34:51.9000695Z * [new tag] ciflow/slow/1732077 -> ciflow/slow/1732077 2025-07-17T08:34:51.9003174Z * [new tag] ciflow/slow/187eb7c -> ciflow/slow/187eb7c 2025-07-17T08:34:51.9005648Z * [new tag] ciflow/slow/1faef89 -> ciflow/slow/1faef89 2025-07-17T08:34:51.9008061Z * [new tag] ciflow/slow/3920ec1 -> ciflow/slow/3920ec1 2025-07-17T08:34:51.9010479Z * [new tag] ciflow/slow/3b7c6b2 -> ciflow/slow/3b7c6b2 2025-07-17T08:34:51.9012934Z * [new tag] ciflow/slow/59a3759 -> ciflow/slow/59a3759 2025-07-17T08:34:51.9015331Z * [new tag] ciflow/slow/70ef0bb -> ciflow/slow/70ef0bb 2025-07-17T08:34:51.9017836Z * [new tag] ciflow/slow/788ff06 -> ciflow/slow/788ff06 2025-07-17T08:34:51.9020538Z * [new tag] ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym -> ciflow/slow/8751002215790a3a88750faa8f4366933e296693-sdym 2025-07-17T08:34:51.9022746Z * [new tag] ciflow/slow/9d85864 -> ciflow/slow/9d85864 2025-07-17T08:34:51.9025130Z * [new tag] ciflow/slow/9ffad5b -> ciflow/slow/9ffad5b 2025-07-17T08:34:51.9027565Z * [new tag] ciflow/slow/a206e8b -> ciflow/slow/a206e8b 2025-07-17T08:34:51.9029990Z * [new tag] ciflow/slow/a837609 -> ciflow/slow/a837609 2025-07-17T08:34:51.9032433Z * [new tag] ciflow/slow/af841f3 -> ciflow/slow/af841f3 2025-07-17T08:34:51.9035275Z * [new tag] ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym -> ciflow/slow/da3aba1e46157c4df504b067477cdf2b3c96b194-sdym 2025-07-17T08:34:51.9037842Z * [new tag] ciflow/triton_binaries/158408 -> ciflow/triton_binaries/158408 2025-07-17T08:34:51.9039918Z * [new tag] ciflow/triton_binaries/158459 -> ciflow/triton_binaries/158459 2025-07-17T08:34:51.9042488Z * [new tag] ciflow/trunk/113258 -> ciflow/trunk/113258 2025-07-17T08:34:51.9044606Z * [new tag] ciflow/trunk/137400 -> ciflow/trunk/137400 2025-07-17T08:34:51.9046717Z * [new tag] ciflow/trunk/139971 -> ciflow/trunk/139971 2025-07-17T08:34:51.9048871Z * [new tag] ciflow/trunk/143712 -> ciflow/trunk/143712 2025-07-17T08:34:51.9051027Z * [new tag] ciflow/trunk/144557 -> ciflow/trunk/144557 2025-07-17T08:34:51.9053146Z * [new tag] ciflow/trunk/147470 -> ciflow/trunk/147470 2025-07-17T08:34:51.9055380Z * [new tag] ciflow/trunk/148180 -> ciflow/trunk/148180 2025-07-17T08:34:51.9057547Z * [new tag] ciflow/trunk/148328 -> ciflow/trunk/148328 2025-07-17T08:34:51.9059650Z * [new tag] ciflow/trunk/148492 -> ciflow/trunk/148492 2025-07-17T08:34:51.9061783Z * [new tag] ciflow/trunk/149003 -> ciflow/trunk/149003 2025-07-17T08:34:51.9063913Z * [new tag] ciflow/trunk/149601 -> ciflow/trunk/149601 2025-07-17T08:34:51.9066025Z * [new tag] ciflow/trunk/149961 -> ciflow/trunk/149961 2025-07-17T08:34:51.9068145Z * [new tag] ciflow/trunk/150312 -> ciflow/trunk/150312 2025-07-17T08:34:51.9070287Z * [new tag] ciflow/trunk/150691 -> ciflow/trunk/150691 2025-07-17T08:34:51.9072465Z * [new tag] ciflow/trunk/150762 -> ciflow/trunk/150762 2025-07-17T08:34:51.9074875Z * [new tag] ciflow/trunk/151777 -> ciflow/trunk/151777 2025-07-17T08:34:51.9076932Z * [new tag] ciflow/trunk/151845 -> ciflow/trunk/151845 2025-07-17T08:34:51.9079016Z * [new tag] ciflow/trunk/152624 -> ciflow/trunk/152624 2025-07-17T08:34:51.9081312Z * [new tag] ciflow/trunk/153666 -> ciflow/trunk/153666 2025-07-17T08:34:51.9083456Z * [new tag] ciflow/trunk/154149 -> ciflow/trunk/154149 2025-07-17T08:34:51.9085695Z * [new tag] ciflow/trunk/154199 -> ciflow/trunk/154199 2025-07-17T08:34:51.9087845Z * [new tag] ciflow/trunk/154694 -> ciflow/trunk/154694 2025-07-17T08:34:51.9090047Z * [new tag] ciflow/trunk/154983 -> ciflow/trunk/154983 2025-07-17T08:34:51.9092228Z * [new tag] ciflow/trunk/155489 -> ciflow/trunk/155489 2025-07-17T08:34:51.9094375Z * [new tag] ciflow/trunk/155958 -> ciflow/trunk/155958 2025-07-17T08:34:51.9096646Z * [new tag] ciflow/trunk/156049 -> ciflow/trunk/156049 2025-07-17T08:34:51.9098907Z * [new tag] ciflow/trunk/156097 -> ciflow/trunk/156097 2025-07-17T08:34:51.9101046Z * [new tag] ciflow/trunk/156141 -> ciflow/trunk/156141 2025-07-17T08:34:51.9103153Z * [new tag] ciflow/trunk/156165 -> ciflow/trunk/156165 2025-07-17T08:34:51.9105313Z * [new tag] ciflow/trunk/156175 -> ciflow/trunk/156175 2025-07-17T08:34:51.9107502Z * [new tag] ciflow/trunk/156192 -> ciflow/trunk/156192 2025-07-17T08:34:51.9109635Z * [new tag] ciflow/trunk/156296 -> ciflow/trunk/156296 2025-07-17T08:34:51.9111770Z * [new tag] ciflow/trunk/156370 -> ciflow/trunk/156370 2025-07-17T08:34:51.9113943Z * [new tag] ciflow/trunk/156559 -> ciflow/trunk/156559 2025-07-17T08:34:51.9116194Z * [new tag] ciflow/trunk/156605 -> ciflow/trunk/156605 2025-07-17T08:34:51.9118316Z * [new tag] ciflow/trunk/156626 -> ciflow/trunk/156626 2025-07-17T08:34:51.9121246Z * [new tag] ciflow/trunk/156666 -> ciflow/trunk/156666 2025-07-17T08:34:51.9123527Z * [new tag] ciflow/trunk/156753 -> ciflow/trunk/156753 2025-07-17T08:34:51.9125740Z * [new tag] ciflow/trunk/156781 -> ciflow/trunk/156781 2025-07-17T08:34:51.9127886Z * [new tag] ciflow/trunk/156874 -> ciflow/trunk/156874 2025-07-17T08:34:51.9130025Z * [new tag] ciflow/trunk/157199 -> ciflow/trunk/157199 2025-07-17T08:34:51.9132230Z * [new tag] ciflow/trunk/157432 -> ciflow/trunk/157432 2025-07-17T08:34:51.9134329Z * [new tag] ciflow/trunk/157520 -> ciflow/trunk/157520 2025-07-17T08:34:51.9136472Z * [new tag] ciflow/trunk/157550 -> ciflow/trunk/157550 2025-07-17T08:34:51.9138607Z * [new tag] ciflow/trunk/157552 -> ciflow/trunk/157552 2025-07-17T08:34:51.9140728Z * [new tag] ciflow/trunk/157580 -> ciflow/trunk/157580 2025-07-17T08:34:51.9142897Z * [new tag] ciflow/trunk/157685 -> ciflow/trunk/157685 2025-07-17T08:34:51.9145026Z * [new tag] ciflow/trunk/157689 -> ciflow/trunk/157689 2025-07-17T08:34:51.9147215Z * [new tag] ciflow/trunk/157699 -> ciflow/trunk/157699 2025-07-17T08:34:51.9149385Z * [new tag] ciflow/trunk/157748 -> ciflow/trunk/157748 2025-07-17T08:34:51.9151484Z * [new tag] ciflow/trunk/157791 -> ciflow/trunk/157791 2025-07-17T08:34:51.9153646Z * [new tag] ciflow/trunk/157804 -> ciflow/trunk/157804 2025-07-17T08:34:51.9156047Z * [new tag] ciflow/trunk/157887 -> ciflow/trunk/157887 2025-07-17T08:34:51.9158356Z * [new tag] ciflow/trunk/157908 -> ciflow/trunk/157908 2025-07-17T08:34:51.9160365Z * [new tag] ciflow/trunk/157910 -> ciflow/trunk/157910 2025-07-17T08:34:51.9162697Z * [new tag] ciflow/trunk/157935 -> ciflow/trunk/157935 2025-07-17T08:34:51.9165011Z * [new tag] ciflow/trunk/157963 -> ciflow/trunk/157963 2025-07-17T08:34:51.9167166Z * [new tag] ciflow/trunk/157994 -> ciflow/trunk/157994 2025-07-17T08:34:51.9169287Z * [new tag] ciflow/trunk/158015 -> ciflow/trunk/158015 2025-07-17T08:34:51.9171410Z * [new tag] ciflow/trunk/158062 -> ciflow/trunk/158062 2025-07-17T08:34:51.9173547Z * [new tag] ciflow/trunk/158072 -> ciflow/trunk/158072 2025-07-17T08:34:51.9175705Z * [new tag] ciflow/trunk/158091 -> ciflow/trunk/158091 2025-07-17T08:34:51.9177862Z * [new tag] ciflow/trunk/158104 -> ciflow/trunk/158104 2025-07-17T08:34:51.9179976Z * [new tag] ciflow/trunk/158119 -> ciflow/trunk/158119 2025-07-17T08:34:51.9182104Z * [new tag] ciflow/trunk/158148 -> ciflow/trunk/158148 2025-07-17T08:34:51.9184376Z * [new tag] ciflow/trunk/158185 -> ciflow/trunk/158185 2025-07-17T08:34:51.9186806Z * [new tag] ciflow/trunk/158203 -> ciflow/trunk/158203 2025-07-17T08:34:51.9188985Z * [new tag] ciflow/trunk/158219 -> ciflow/trunk/158219 2025-07-17T08:34:51.9191087Z * [new tag] ciflow/trunk/158220 -> ciflow/trunk/158220 2025-07-17T08:34:51.9193263Z * [new tag] ciflow/trunk/158222 -> ciflow/trunk/158222 2025-07-17T08:34:51.9195695Z * [new tag] ciflow/trunk/158223 -> ciflow/trunk/158223 2025-07-17T08:34:51.9197855Z * [new tag] ciflow/trunk/158224 -> ciflow/trunk/158224 2025-07-17T08:34:51.9200270Z * [new tag] ciflow/trunk/158229 -> ciflow/trunk/158229 2025-07-17T08:34:51.9202406Z * [new tag] ciflow/trunk/158249 -> ciflow/trunk/158249 2025-07-17T08:34:51.9204529Z * [new tag] ciflow/trunk/158259 -> ciflow/trunk/158259 2025-07-17T08:34:51.9206687Z * [new tag] ciflow/trunk/158289 -> ciflow/trunk/158289 2025-07-17T08:34:51.9208842Z * [new tag] ciflow/trunk/158312 -> ciflow/trunk/158312 2025-07-17T08:34:51.9211146Z * [new tag] ciflow/trunk/158323 -> ciflow/trunk/158323 2025-07-17T08:34:51.9213343Z * [new tag] ciflow/trunk/158356 -> ciflow/trunk/158356 2025-07-17T08:34:51.9215553Z * [new tag] ciflow/trunk/158363 -> ciflow/trunk/158363 2025-07-17T08:34:51.9217714Z * [new tag] ciflow/trunk/158368 -> ciflow/trunk/158368 2025-07-17T08:34:51.9220020Z * [new tag] ciflow/trunk/158373 -> ciflow/trunk/158373 2025-07-17T08:34:51.9222124Z * [new tag] ciflow/trunk/158377 -> ciflow/trunk/158377 2025-07-17T08:34:51.9224402Z * [new tag] ciflow/trunk/158380 -> ciflow/trunk/158380 2025-07-17T08:34:51.9226659Z * [new tag] ciflow/trunk/158400 -> ciflow/trunk/158400 2025-07-17T08:34:51.9228824Z * [new tag] ciflow/trunk/158418 -> ciflow/trunk/158418 2025-07-17T08:34:51.9231023Z * [new tag] ciflow/trunk/158424 -> ciflow/trunk/158424 2025-07-17T08:34:51.9233132Z * [new tag] ciflow/trunk/158430 -> ciflow/trunk/158430 2025-07-17T08:34:51.9237352Z * [new tag] ciflow/trunk/158442 -> ciflow/trunk/158442 2025-07-17T08:34:51.9239653Z * [new tag] ciflow/trunk/158458 -> ciflow/trunk/158458 2025-07-17T08:34:51.9241825Z * [new tag] ciflow/trunk/158459 -> ciflow/trunk/158459 2025-07-17T08:34:51.9244302Z * [new tag] ciflow/trunk/158473 -> ciflow/trunk/158473 2025-07-17T08:34:51.9246308Z * [new tag] ciflow/trunk/158479 -> ciflow/trunk/158479 2025-07-17T08:34:51.9248464Z * [new tag] ciflow/trunk/158485 -> ciflow/trunk/158485 2025-07-17T08:34:51.9250686Z * [new tag] ciflow/trunk/158492 -> ciflow/trunk/158492 2025-07-17T08:34:51.9252864Z * [new tag] ciflow/trunk/158524 -> ciflow/trunk/158524 2025-07-17T08:34:51.9255194Z * [new tag] ciflow/trunk/158532 -> ciflow/trunk/158532 2025-07-17T08:34:51.9257360Z * [new tag] ciflow/trunk/158541 -> ciflow/trunk/158541 2025-07-17T08:34:51.9259517Z * [new tag] ciflow/trunk/158543 -> ciflow/trunk/158543 2025-07-17T08:34:51.9262577Z * [new tag] ciflow/unstable/123 -> ciflow/unstable/123 2025-07-17T08:34:51.9265317Z * [new tag] ciflow/win-arm64/157935 -> ciflow/win-arm64/157935 2025-07-17T08:34:51.9267996Z * [new tag] ciflow/xpu/139971 -> ciflow/xpu/139971 2025-07-17T08:34:51.9270161Z * [new tag] ciflow/xpu/150218 -> ciflow/xpu/150218 2025-07-17T08:34:51.9272474Z * [new tag] ciflow/xpu/155200 -> ciflow/xpu/155200 2025-07-17T08:34:51.9274788Z * [new tag] ciflow/xpu/156812 -> ciflow/xpu/156812 2025-07-17T08:34:51.9277050Z * [new tag] ciflow/xpu/157699 -> ciflow/xpu/157699 2025-07-17T08:34:51.9279234Z * [new tag] ciflow/xpu/157954 -> ciflow/xpu/157954 2025-07-17T08:34:51.9281470Z * [new tag] ciflow/xpu/158336 -> ciflow/xpu/158336 2025-07-17T08:34:51.9283879Z * [new tag] ciflow/xpu/158337 -> ciflow/xpu/158337 2025-07-17T08:34:51.9286152Z * [new tag] ciflow/xpu/158340 -> ciflow/xpu/158340 2025-07-17T08:34:51.9288976Z * [new tag] ciflow/xpu/158533 -> ciflow/xpu/158533 2025-07-17T08:34:51.9290640Z * [new tag] ciflow/xpu/158542 -> ciflow/xpu/158542 2025-07-17T08:34:51.9293037Z * [new tag] cslpull75 -> cslpull75 2025-07-17T08:34:51.9295390Z * [new tag] cslpull76 -> cslpull76 2025-07-17T08:34:51.9297718Z * [new tag] cslpull77 -> cslpull77 2025-07-17T08:34:51.9299965Z * [new tag] cslpull78 -> cslpull78 2025-07-17T08:34:51.9302219Z * [new tag] cslpull79 -> cslpull79 2025-07-17T08:34:51.9304548Z * [new tag] cslpull80 -> cslpull80 2025-07-17T08:34:51.9306865Z * [new tag] cslpull81 -> cslpull81 2025-07-17T08:34:51.9309141Z * [new tag] cslpull82 -> cslpull82 2025-07-17T08:34:51.9311541Z * [new tag] cslpull83 -> cslpull83 2025-07-17T08:34:51.9313852Z * [new tag] cslpull84 -> cslpull84 2025-07-17T08:34:51.9316314Z * [new tag] cslpull85 -> cslpull85 2025-07-17T08:34:51.9318674Z * [new tag] cslpull86 -> cslpull86 2025-07-17T08:34:51.9320979Z * [new tag] cslpull87 -> cslpull87 2025-07-17T08:34:51.9323334Z * [new tag] cslpull88 -> cslpull88 2025-07-17T08:34:51.9325714Z * [new tag] cslpull89 -> cslpull89 2025-07-17T08:34:51.9327866Z * [new tag] cslpull90 -> cslpull90 2025-07-17T08:34:51.9330611Z * [new tag] cslpull91 -> cslpull91 2025-07-17T08:34:51.9332944Z * [new tag] cslpull92 -> cslpull92 2025-07-17T08:34:51.9335450Z * [new tag] flight_5 -> flight_5 2025-07-17T08:34:51.9337706Z * [new tag] flight_5.1 -> flight_5.1 2025-07-17T08:34:51.9340098Z * [new tag] flight_5.2 -> flight_5.2 2025-07-17T08:34:51.9342341Z * [new tag] flight_5.3 -> flight_5.3 2025-07-17T08:34:51.9344679Z * [new tag] forpull1 -> forpull1 2025-07-17T08:34:51.9347600Z * [new tag] malfet/tag-2ef5611 -> malfet/tag-2ef5611 2025-07-17T08:34:51.9349846Z * [new tag] malfet/tag-317b1a0 -> malfet/tag-317b1a0 2025-07-17T08:34:51.9352097Z * [new tag] malfet/tag-ec6f767 -> malfet/tag-ec6f767 2025-07-17T08:34:51.9354676Z * [new tag] nightly-binary -> nightly-binary 2025-07-17T08:34:51.9356944Z * [new tag] sqzhang_flight4_plus -> sqzhang_flight4_plus 2025-07-17T08:34:51.9359266Z * [new tag] sqzhang_flight_3 -> sqzhang_flight_3 2025-07-17T08:34:51.9362213Z * [new tag] trunk/0083032e7559dc8f02483ba60373adfcdaf9dae6 -> trunk/0083032e7559dc8f02483ba60373adfcdaf9dae6 2025-07-17T08:34:51.9364548Z * [new tag] trunk/008345be9d0c32f67459bcf3e6705be43d496f74 -> trunk/008345be9d0c32f67459bcf3e6705be43d496f74 2025-07-17T08:34:51.9366854Z * [new tag] trunk/00ae620b9f72cc751672c2fd92a04fc86704a87a -> trunk/00ae620b9f72cc751672c2fd92a04fc86704a87a 2025-07-17T08:34:51.9369007Z * [new tag] trunk/0105cd89ab508ec56126c1de85c8f5b5acc446b5 -> trunk/0105cd89ab508ec56126c1de85c8f5b5acc446b5 2025-07-17T08:34:51.9371263Z * [new tag] trunk/011026205a9d4c38458130f8ca242028f6184bf0 -> trunk/011026205a9d4c38458130f8ca242028f6184bf0 2025-07-17T08:34:51.9373514Z * [new tag] trunk/013cf1e3302d27de36588cf7a7130d76a5686bad -> trunk/013cf1e3302d27de36588cf7a7130d76a5686bad 2025-07-17T08:34:51.9375852Z * [new tag] trunk/013dfeabb405274546f19637c04f8a5f75923316 -> trunk/013dfeabb405274546f19637c04f8a5f75923316 2025-07-17T08:34:51.9378161Z * [new tag] trunk/018e9826a2ed09cfcf7424e0b3215bef5ad6499e -> trunk/018e9826a2ed09cfcf7424e0b3215bef5ad6499e 2025-07-17T08:34:51.9380373Z * [new tag] trunk/019e30e3b80d091d64253df4cdd149713e3e910e -> trunk/019e30e3b80d091d64253df4cdd149713e3e910e 2025-07-17T08:34:51.9382676Z * [new tag] trunk/01b0f09931d47bd2716398a0c335b2807dc3074d -> trunk/01b0f09931d47bd2716398a0c335b2807dc3074d 2025-07-17T08:34:51.9385062Z * [new tag] trunk/01b8f5e685fdb34aa50b73a84a55fc285e6b904a -> trunk/01b8f5e685fdb34aa50b73a84a55fc285e6b904a 2025-07-17T08:34:51.9387254Z * [new tag] trunk/02080c2cd94b37ad92e3d8803bc773f6abd2ae2f -> trunk/02080c2cd94b37ad92e3d8803bc773f6abd2ae2f 2025-07-17T08:34:51.9389498Z * [new tag] trunk/020da744370f6ee23e377357e9acc330b5610a67 -> trunk/020da744370f6ee23e377357e9acc330b5610a67 2025-07-17T08:34:51.9391763Z * [new tag] trunk/023887fc5af5a7273d7eb0388fdf1f0e87eafb28 -> trunk/023887fc5af5a7273d7eb0388fdf1f0e87eafb28 2025-07-17T08:34:51.9394155Z * [new tag] trunk/02608e560a1d7b669ad450a89542f7e9ba068658 -> trunk/02608e560a1d7b669ad450a89542f7e9ba068658 2025-07-17T08:34:51.9396582Z * [new tag] trunk/02715d0876bc66f46b9c7b4f277bedf8e251cc82 -> trunk/02715d0876bc66f46b9c7b4f277bedf8e251cc82 2025-07-17T08:34:51.9399113Z * [new tag] trunk/02724b5f649b93ef7960962bdde7a667c0893d21 -> trunk/02724b5f649b93ef7960962bdde7a667c0893d21 2025-07-17T08:34:51.9401469Z * [new tag] trunk/029e2b05c225588098d3eba445fd04189691f77d -> trunk/029e2b05c225588098d3eba445fd04189691f77d 2025-07-17T08:34:51.9403713Z * [new tag] trunk/02a9d9095f397f56fe64ef07ced8e7ffed6dcba1 -> trunk/02a9d9095f397f56fe64ef07ced8e7ffed6dcba1 2025-07-17T08:34:51.9406167Z * [new tag] trunk/02c7ab2f9baac05bd199392b70bc016d55f99b13 -> trunk/02c7ab2f9baac05bd199392b70bc016d55f99b13 2025-07-17T08:34:51.9408112Z * [new tag] trunk/03023f178c611249d41c62369ba95fe54316fe90 -> trunk/03023f178c611249d41c62369ba95fe54316fe90 2025-07-17T08:34:51.9410445Z * [new tag] trunk/03488d820c292b8ec4bfd9a4e25d5f28068c9375 -> trunk/03488d820c292b8ec4bfd9a4e25d5f28068c9375 2025-07-17T08:34:51.9412784Z * [new tag] trunk/034a7f6437bec853a938dbc19b549b9319b67f69 -> trunk/034a7f6437bec853a938dbc19b549b9319b67f69 2025-07-17T08:34:51.9414976Z * [new tag] trunk/0364db7cd14ffa67b48ef8c27fefbb3eed2b065d -> trunk/0364db7cd14ffa67b48ef8c27fefbb3eed2b065d 2025-07-17T08:34:51.9417208Z * [new tag] trunk/03852ddc22350eb8b6ed6b61777639ce6080f3dc -> trunk/03852ddc22350eb8b6ed6b61777639ce6080f3dc 2025-07-17T08:34:51.9419492Z * [new tag] trunk/039a1ce0ebbeb0adacbf9537c4085d977dba150a -> trunk/039a1ce0ebbeb0adacbf9537c4085d977dba150a 2025-07-17T08:34:51.9421876Z * [new tag] trunk/03b307575a98dc1d953c9d3521a9489e0e61e70c -> trunk/03b307575a98dc1d953c9d3521a9489e0e61e70c 2025-07-17T08:34:51.9424130Z * [new tag] trunk/04178d347c6baec5613f8063a2c77cf6f9657ee2 -> trunk/04178d347c6baec5613f8063a2c77cf6f9657ee2 2025-07-17T08:34:51.9426437Z * [new tag] trunk/04349f9ee541c7d07cc057bbe739f46bd4c30dcc -> trunk/04349f9ee541c7d07cc057bbe739f46bd4c30dcc 2025-07-17T08:34:51.9428731Z * [new tag] trunk/049dc48d1edf49c26d493fdb271fe5e5adc9c985 -> trunk/049dc48d1edf49c26d493fdb271fe5e5adc9c985 2025-07-17T08:34:51.9431160Z * [new tag] trunk/04b91a9e43518fcd34a474a832bac05d4583fcc3 -> trunk/04b91a9e43518fcd34a474a832bac05d4583fcc3 2025-07-17T08:34:51.9433408Z * [new tag] trunk/04bd7e6850e8efec77994963ffee87549555b9c3 -> trunk/04bd7e6850e8efec77994963ffee87549555b9c3 2025-07-17T08:34:51.9435862Z * [new tag] trunk/04cf2c9d24b7d9e7f7cdd4cf98c8bd0908c38110 -> trunk/04cf2c9d24b7d9e7f7cdd4cf98c8bd0908c38110 2025-07-17T08:34:51.9438269Z * [new tag] trunk/0504480f37714a289b2ba32c9cf32a5e50e86d38 -> trunk/0504480f37714a289b2ba32c9cf32a5e50e86d38 2025-07-17T08:34:51.9440563Z * [new tag] trunk/054cd4ca28d17507df46054fe355c697f6a07ce8 -> trunk/054cd4ca28d17507df46054fe355c697f6a07ce8 2025-07-17T08:34:51.9442918Z * [new tag] trunk/058fb1790f2c474cd4ecb5ec625eef896c554544 -> trunk/058fb1790f2c474cd4ecb5ec625eef896c554544 2025-07-17T08:34:51.9445164Z * [new tag] trunk/0596323c35d245099a4f4d3172b0efa20a639c84 -> trunk/0596323c35d245099a4f4d3172b0efa20a639c84 2025-07-17T08:34:51.9447430Z * [new tag] trunk/05d7288e316ae5c9c661c4529f9f130a46263e5b -> trunk/05d7288e316ae5c9c661c4529f9f130a46263e5b 2025-07-17T08:34:51.9449862Z * [new tag] trunk/05dd638ee98b36254c84095894c36fd0e7d95544 -> trunk/05dd638ee98b36254c84095894c36fd0e7d95544 2025-07-17T08:34:51.9452157Z * [new tag] trunk/05dfd312cfbfdecc6cb1e7d1d0bb4ee18370ae7e -> trunk/05dfd312cfbfdecc6cb1e7d1d0bb4ee18370ae7e 2025-07-17T08:34:51.9454253Z * [new tag] trunk/05faba40287cf7d8734da96cb2e904f39710bf29 -> trunk/05faba40287cf7d8734da96cb2e904f39710bf29 2025-07-17T08:34:51.9457433Z * [new tag] trunk/060838c2312ad207c7afe2c86f8a484afea5f328 -> trunk/060838c2312ad207c7afe2c86f8a484afea5f328 2025-07-17T08:34:51.9459676Z * [new tag] trunk/0629dfb860b00a10550d91dd395968f663f45fdd -> trunk/0629dfb860b00a10550d91dd395968f663f45fdd 2025-07-17T08:34:51.9461958Z * [new tag] trunk/06408dae49d06b6146fdd9d7a37eb5dde4f5e78d -> trunk/06408dae49d06b6146fdd9d7a37eb5dde4f5e78d 2025-07-17T08:34:51.9464251Z * [new tag] trunk/0640cfa38c1426a41ab4a0b3e3dab7c730cdc2ad -> trunk/0640cfa38c1426a41ab4a0b3e3dab7c730cdc2ad 2025-07-17T08:34:51.9466656Z * [new tag] trunk/064288cbab94c9931ca2296a2b9723e864f9050a -> trunk/064288cbab94c9931ca2296a2b9723e864f9050a 2025-07-17T08:34:51.9468797Z * [new tag] trunk/064a7db7fc3b06eb8e14d6c7d767f87fcf396ab6 -> trunk/064a7db7fc3b06eb8e14d6c7d767f87fcf396ab6 2025-07-17T08:34:51.9471051Z * [new tag] trunk/067fd0b3abd28f1b7dae883ffec5bc7459d35970 -> trunk/067fd0b3abd28f1b7dae883ffec5bc7459d35970 2025-07-17T08:34:51.9473397Z * [new tag] trunk/06930706a195144cd6172a5f2eb505cec56e9ec3 -> trunk/06930706a195144cd6172a5f2eb505cec56e9ec3 2025-07-17T08:34:51.9476187Z * [new tag] trunk/06a40b685056f60c4fbdc10d09a85b9fc3fca34f -> trunk/06a40b685056f60c4fbdc10d09a85b9fc3fca34f 2025-07-17T08:34:51.9478462Z * [new tag] trunk/06a67a8948dac9d02f22b3e2591a43b60981cdb4 -> trunk/06a67a8948dac9d02f22b3e2591a43b60981cdb4 2025-07-17T08:34:51.9480735Z * [new tag] trunk/06b3265cb1a27ea3090dc1fc86a6ce17518f989c -> trunk/06b3265cb1a27ea3090dc1fc86a6ce17518f989c 2025-07-17T08:34:51.9483014Z * [new tag] trunk/06f39a71b6cd14e0844b5eb1e0e085f7ae78d221 -> trunk/06f39a71b6cd14e0844b5eb1e0e085f7ae78d221 2025-07-17T08:34:51.9485110Z * [new tag] trunk/070aa59e491aa6987459cd1941ef647f2ad13858 -> trunk/070aa59e491aa6987459cd1941ef647f2ad13858 2025-07-17T08:34:51.9487503Z * [new tag] trunk/070e580d301efd626e327d1f88e0921b476d0f30 -> trunk/070e580d301efd626e327d1f88e0921b476d0f30 2025-07-17T08:34:51.9489784Z * [new tag] trunk/0756ebcd4890b19072e586fb2cbd2b25a80f91f4 -> trunk/0756ebcd4890b19072e586fb2cbd2b25a80f91f4 2025-07-17T08:34:51.9492077Z * [new tag] trunk/0797b2b6a80cf70a7accc3d5413186e7693d4451 -> trunk/0797b2b6a80cf70a7accc3d5413186e7693d4451 2025-07-17T08:34:51.9494295Z * [new tag] trunk/07bb097698223fe42d54b019638dd026c75f09a9 -> trunk/07bb097698223fe42d54b019638dd026c75f09a9 2025-07-17T08:34:51.9496536Z * [new tag] trunk/07da8a469bdd69966e1f832e3d09d991798f854c -> trunk/07da8a469bdd69966e1f832e3d09d991798f854c 2025-07-17T08:34:51.9498900Z * [new tag] trunk/07e340e29ccbe18ef59e8f60ec3f03ccc44e5e9c -> trunk/07e340e29ccbe18ef59e8f60ec3f03ccc44e5e9c 2025-07-17T08:34:51.9501373Z * [new tag] trunk/07eb374e7eefaa84a997d0adaaeecb4706d60e74 -> trunk/07eb374e7eefaa84a997d0adaaeecb4706d60e74 2025-07-17T08:34:51.9503697Z * [new tag] trunk/08274640028f68418d8f889d01af577b90ac2ec1 -> trunk/08274640028f68418d8f889d01af577b90ac2ec1 2025-07-17T08:34:51.9506102Z * [new tag] trunk/085f270a00b4452bbb005d6b3d448e9d0b9d6fa0 -> trunk/085f270a00b4452bbb005d6b3d448e9d0b9d6fa0 2025-07-17T08:34:51.9508347Z * [new tag] trunk/0860606729d16b9855157404da9b0d4654f3c626 -> trunk/0860606729d16b9855157404da9b0d4654f3c626 2025-07-17T08:34:51.9510537Z * [new tag] trunk/08799217aeb17128d89d675ce5b537761286417a -> trunk/08799217aeb17128d89d675ce5b537761286417a 2025-07-17T08:34:51.9512852Z * [new tag] trunk/08d15d3ec15b9b3fce81713cde20901806ecef42 -> trunk/08d15d3ec15b9b3fce81713cde20901806ecef42 2025-07-17T08:34:51.9515112Z * [new tag] trunk/08dae945ae380d80efbaf140a95abfc5d96e5100 -> trunk/08dae945ae380d80efbaf140a95abfc5d96e5100 2025-07-17T08:34:51.9517578Z * [new tag] trunk/08e9dd280f497fc820e35c458c843dc44f0282c6 -> trunk/08e9dd280f497fc820e35c458c843dc44f0282c6 2025-07-17T08:34:51.9519816Z * [new tag] trunk/092aed1b18c31042d7772a6f9ec1959df4bf5c3c -> trunk/092aed1b18c31042d7772a6f9ec1959df4bf5c3c 2025-07-17T08:34:51.9522091Z * [new tag] trunk/09328eb02f5412d2211b5fd638ce82d0e03b9c1f -> trunk/09328eb02f5412d2211b5fd638ce82d0e03b9c1f 2025-07-17T08:34:51.9524488Z * [new tag] trunk/0935a97d959918727700ae2c1ebeadc5b12a2b6c -> trunk/0935a97d959918727700ae2c1ebeadc5b12a2b6c 2025-07-17T08:34:51.9526616Z * [new tag] trunk/093aaccae254355b3637ef8760e762a3f9c3783c -> trunk/093aaccae254355b3637ef8760e762a3f9c3783c 2025-07-17T08:34:51.9528997Z * [new tag] trunk/093fd47dbef4f08691210de6c08cf269f28e4de5 -> trunk/093fd47dbef4f08691210de6c08cf269f28e4de5 2025-07-17T08:34:51.9531104Z * [new tag] trunk/099d0d6121125062ebc05771c8330cb7cd8d053a -> trunk/099d0d6121125062ebc05771c8330cb7cd8d053a 2025-07-17T08:34:51.9533403Z * [new tag] trunk/09db3a22e8783c4841697317688ba9467c7cc457 -> trunk/09db3a22e8783c4841697317688ba9467c7cc457 2025-07-17T08:34:51.9535749Z * [new tag] trunk/09ffba3cf79eeedf48a0fb39a26dcc0294872480 -> trunk/09ffba3cf79eeedf48a0fb39a26dcc0294872480 2025-07-17T08:34:51.9538044Z * [new tag] trunk/0a0023d9840570938a10472418eaac85098ec41b -> trunk/0a0023d9840570938a10472418eaac85098ec41b 2025-07-17T08:34:51.9540335Z * [new tag] trunk/0a092c7de67767eed3adb7abc0d49dfd3d26d0cd -> trunk/0a092c7de67767eed3adb7abc0d49dfd3d26d0cd 2025-07-17T08:34:51.9542775Z * [new tag] trunk/0a16818d5b3fdf0fb8148dd6b849687251a56376 -> trunk/0a16818d5b3fdf0fb8148dd6b849687251a56376 2025-07-17T08:34:51.9545043Z * [new tag] trunk/0a2ec7681d2af973d8daaf7905431a088739dc90 -> trunk/0a2ec7681d2af973d8daaf7905431a088739dc90 2025-07-17T08:34:51.9547293Z * [new tag] trunk/0a624c2dc5675e35917976467a8c533baf2a98a0 -> trunk/0a624c2dc5675e35917976467a8c533baf2a98a0 2025-07-17T08:34:51.9549651Z * [new tag] trunk/0a63053fe90405cc6ee5e42fa007e32ed1e2cbac -> trunk/0a63053fe90405cc6ee5e42fa007e32ed1e2cbac 2025-07-17T08:34:51.9551800Z * [new tag] trunk/0a6b66c881cba3f6a6c1a3cb8ddf698846d99822 -> trunk/0a6b66c881cba3f6a6c1a3cb8ddf698846d99822 2025-07-17T08:34:51.9554227Z * [new tag] trunk/0a99b026d6bd0f67dc2c0a20fe3228ddc4144854 -> trunk/0a99b026d6bd0f67dc2c0a20fe3228ddc4144854 2025-07-17T08:34:51.9556634Z * [new tag] trunk/0a9d450168ce58b2bb7f2cedc27a61012123564f -> trunk/0a9d450168ce58b2bb7f2cedc27a61012123564f 2025-07-17T08:34:51.9558841Z * [new tag] trunk/0ab075a69e4577a60c4dcbff7bcc2ecd0a15ce46 -> trunk/0ab075a69e4577a60c4dcbff7bcc2ecd0a15ce46 2025-07-17T08:34:51.9561178Z * [new tag] trunk/0ad88a2224e9f0a2ed4f405e827a745945d505ce -> trunk/0ad88a2224e9f0a2ed4f405e827a745945d505ce 2025-07-17T08:34:51.9563479Z * [new tag] trunk/0aed855b2bde6d9bd045bb20cc24544a9f2fb72b -> trunk/0aed855b2bde6d9bd045bb20cc24544a9f2fb72b 2025-07-17T08:34:51.9565754Z * [new tag] trunk/0b19d463d963a0b2ee5558d2c0bb79b2cbff6e64 -> trunk/0b19d463d963a0b2ee5558d2c0bb79b2cbff6e64 2025-07-17T08:34:51.9568064Z * [new tag] trunk/0b62465b99b23cb4afcd07424676cce34a676041 -> trunk/0b62465b99b23cb4afcd07424676cce34a676041 2025-07-17T08:34:51.9570376Z * [new tag] trunk/0b677560e686d828f9f5ed1cb87b1d7a54ee2781 -> trunk/0b677560e686d828f9f5ed1cb87b1d7a54ee2781 2025-07-17T08:34:51.9572646Z * [new tag] trunk/0b6c0898e6c352c8ea93daec854e704b41485375 -> trunk/0b6c0898e6c352c8ea93daec854e704b41485375 2025-07-17T08:34:51.9574913Z * [new tag] trunk/0bb733ba230051301b3fb3fa49d1d6662744b395 -> trunk/0bb733ba230051301b3fb3fa49d1d6662744b395 2025-07-17T08:34:51.9577173Z * [new tag] trunk/0bce39026917d10995ff47352c3da9fff0ec31ff -> trunk/0bce39026917d10995ff47352c3da9fff0ec31ff 2025-07-17T08:34:51.9579704Z * [new tag] trunk/0c58bdd8fb5f269aef100af8e2c43cfcf5f1f9dd -> trunk/0c58bdd8fb5f269aef100af8e2c43cfcf5f1f9dd 2025-07-17T08:34:51.9582096Z * [new tag] trunk/0ca2a79f5b80c1ec8be95e6f7de4182dd90f3502 -> trunk/0ca2a79f5b80c1ec8be95e6f7de4182dd90f3502 2025-07-17T08:34:51.9584396Z * [new tag] trunk/0cb36e2d62c811fcddea4c6d28b1c65246cdd160 -> trunk/0cb36e2d62c811fcddea4c6d28b1c65246cdd160 2025-07-17T08:34:51.9594090Z * [new tag] trunk/0cb85c188f3665c23766cda25c8169f1a436d36b -> trunk/0cb85c188f3665c23766cda25c8169f1a436d36b 2025-07-17T08:34:51.9594859Z * [new tag] trunk/0d01bafc34fc99a0b3e46cbf1ecfd8f97563efa6 -> trunk/0d01bafc34fc99a0b3e46cbf1ecfd8f97563efa6 2025-07-17T08:34:51.9595399Z * [new tag] trunk/0d17029fea3d96bb88b19912946648b47f8e003d -> trunk/0d17029fea3d96bb88b19912946648b47f8e003d 2025-07-17T08:34:51.9595883Z * [new tag] trunk/0d3d84d866040fad1b21a618c44802951a3cb15e -> trunk/0d3d84d866040fad1b21a618c44802951a3cb15e 2025-07-17T08:34:51.9596372Z * [new tag] trunk/0d77364ee3ff6435aa93df3b0303db66b5fc3254 -> trunk/0d77364ee3ff6435aa93df3b0303db66b5fc3254 2025-07-17T08:34:51.9598324Z * [new tag] trunk/0d8c029584b61da059b6225537764eadd2fa1223 -> trunk/0d8c029584b61da059b6225537764eadd2fa1223 2025-07-17T08:34:51.9600567Z * [new tag] trunk/0d8e4e2327137b370d91bd23931cd7691a95cc33 -> trunk/0d8e4e2327137b370d91bd23931cd7691a95cc33 2025-07-17T08:34:51.9602933Z * [new tag] trunk/0db3e0cf29604dae1007a678603e4dffd1c57562 -> trunk/0db3e0cf29604dae1007a678603e4dffd1c57562 2025-07-17T08:34:51.9605188Z * [new tag] trunk/0decd966af9cdcb7ab4410cf475d2fc09f2dea0c -> trunk/0decd966af9cdcb7ab4410cf475d2fc09f2dea0c 2025-07-17T08:34:51.9607448Z * [new tag] trunk/0e2013a12da944930012265bd1b75d0a19af4d7c -> trunk/0e2013a12da944930012265bd1b75d0a19af4d7c 2025-07-17T08:34:51.9609727Z * [new tag] trunk/0e47312ae5a687f0aed61db753d03180118cddc4 -> trunk/0e47312ae5a687f0aed61db753d03180118cddc4 2025-07-17T08:34:51.9611977Z * [new tag] trunk/0e9d8032a3431505c8a6a341375af9cb27afd537 -> trunk/0e9d8032a3431505c8a6a341375af9cb27afd537 2025-07-17T08:34:51.9614237Z * [new tag] trunk/0edc1b91f708c7635cf0f286f43724665934dc89 -> trunk/0edc1b91f708c7635cf0f286f43724665934dc89 2025-07-17T08:34:51.9616470Z * [new tag] trunk/0f0c01071476145fca7d70c68d029a88aeefe72c -> trunk/0f0c01071476145fca7d70c68d029a88aeefe72c 2025-07-17T08:34:51.9618747Z * [new tag] trunk/0f21fa84fb605c61482e4218df89f8bb1ef70c14 -> trunk/0f21fa84fb605c61482e4218df89f8bb1ef70c14 2025-07-17T08:34:51.9621005Z * [new tag] trunk/0f31445139d7ffaeeae95a230c536ef97970af5c -> trunk/0f31445139d7ffaeeae95a230c536ef97970af5c 2025-07-17T08:34:51.9623334Z * [new tag] trunk/0f47e76937f092263dd579e71d45e3193fbbc5df -> trunk/0f47e76937f092263dd579e71d45e3193fbbc5df 2025-07-17T08:34:51.9625464Z * [new tag] trunk/0f9c1b374fbb6f3b999855b26ebb869387b33dfb -> trunk/0f9c1b374fbb6f3b999855b26ebb869387b33dfb 2025-07-17T08:34:51.9627806Z * [new tag] trunk/0fa361e4295bbd8ac8b1f1ea74852535326f08f2 -> trunk/0fa361e4295bbd8ac8b1f1ea74852535326f08f2 2025-07-17T08:34:51.9630112Z * [new tag] trunk/0fd711df19a6523c9de6c820640ae53e32de14c6 -> trunk/0fd711df19a6523c9de6c820640ae53e32de14c6 2025-07-17T08:34:51.9632368Z * [new tag] trunk/1036f6d114bc22a9b4cf620cf7f8364ea2fd7a60 -> trunk/1036f6d114bc22a9b4cf620cf7f8364ea2fd7a60 2025-07-17T08:34:51.9635145Z * [new tag] trunk/104493487852f0cc232547ecbf7bebf5c7fe5982 -> trunk/104493487852f0cc232547ecbf7bebf5c7fe5982 2025-07-17T08:34:51.9637367Z * [new tag] trunk/1051b93192710d2e4bdf5fecd8e66c61731b55e1 -> trunk/1051b93192710d2e4bdf5fecd8e66c61731b55e1 2025-07-17T08:34:51.9639668Z * [new tag] trunk/10cd1de5183e4fcd6e316be497904b9c86d4db67 -> trunk/10cd1de5183e4fcd6e316be497904b9c86d4db67 2025-07-17T08:34:51.9641992Z * [new tag] trunk/10cef1e25d39c1ff1ac93036302ce1b008cd4469 -> trunk/10cef1e25d39c1ff1ac93036302ce1b008cd4469 2025-07-17T08:34:51.9644336Z * [new tag] trunk/10d41c7d20747829a4ee5b994bf466071e4c7e32 -> trunk/10d41c7d20747829a4ee5b994bf466071e4c7e32 2025-07-17T08:34:51.9646635Z * [new tag] trunk/10fb98a004e129ea2a2074cc05e5173c056f728c -> trunk/10fb98a004e129ea2a2074cc05e5173c056f728c 2025-07-17T08:34:51.9648776Z * [new tag] trunk/110ae0f4333b289b1ee798ab93447e6ede16300f -> trunk/110ae0f4333b289b1ee798ab93447e6ede16300f 2025-07-17T08:34:51.9651306Z * [new tag] trunk/1155c53e7d8c25dabc618d7f7b324ee97c89f40e -> trunk/1155c53e7d8c25dabc618d7f7b324ee97c89f40e 2025-07-17T08:34:51.9653483Z * [new tag] trunk/117db5601d78cbc746b35eef71fc815e042e903f -> trunk/117db5601d78cbc746b35eef71fc815e042e903f 2025-07-17T08:34:51.9655645Z * [new tag] trunk/11a86ad2fa3be5cf91a7a2b99c4df44017bc92c7 -> trunk/11a86ad2fa3be5cf91a7a2b99c4df44017bc92c7 2025-07-17T08:34:51.9657925Z * [new tag] trunk/11bb1ece50ade57b963afe29b27cab6a0a56ff7d -> trunk/11bb1ece50ade57b963afe29b27cab6a0a56ff7d 2025-07-17T08:34:51.9660123Z * [new tag] trunk/11bc29856d8180571146209cea987ac9397ab444 -> trunk/11bc29856d8180571146209cea987ac9397ab444 2025-07-17T08:34:51.9662388Z * [new tag] trunk/11d6ad8b2e7359e2f654474f31692ba1aec67980 -> trunk/11d6ad8b2e7359e2f654474f31692ba1aec67980 2025-07-17T08:34:51.9664675Z * [new tag] trunk/11f7e2f1459b09b1ece26ea785b6be77daa36923 -> trunk/11f7e2f1459b09b1ece26ea785b6be77daa36923 2025-07-17T08:34:51.9666904Z * [new tag] trunk/12151c96d9202875638ea2c695d5647c38368c46 -> trunk/12151c96d9202875638ea2c695d5647c38368c46 2025-07-17T08:34:51.9669143Z * [new tag] trunk/127695eb5c973f9fdba24c47b465e30a19292582 -> trunk/127695eb5c973f9fdba24c47b465e30a19292582 2025-07-17T08:34:51.9671877Z * [new tag] trunk/12b02137af5cd6743adfe5a0ffe0d9b453cd013e -> trunk/12b02137af5cd6743adfe5a0ffe0d9b453cd013e 2025-07-17T08:34:51.9674176Z * [new tag] trunk/12cb06e574738a15a59ee9384493682fb5bc6d2b -> trunk/12cb06e574738a15a59ee9384493682fb5bc6d2b 2025-07-17T08:34:51.9676573Z * [new tag] trunk/12f9942b107acc9d7acf9591818c826ef972a0f5 -> trunk/12f9942b107acc9d7acf9591818c826ef972a0f5 2025-07-17T08:34:51.9678817Z * [new tag] trunk/130d4973bd036e539e995dd0bc20080c2570b6da -> trunk/130d4973bd036e539e995dd0bc20080c2570b6da 2025-07-17T08:34:51.9681104Z * [new tag] trunk/132babe7e0b668d31e6efdf339687d48b2ccedc7 -> trunk/132babe7e0b668d31e6efdf339687d48b2ccedc7 2025-07-17T08:34:51.9683359Z * [new tag] trunk/1339e88105cf72586d728d4f7e0d71b81e08d933 -> trunk/1339e88105cf72586d728d4f7e0d71b81e08d933 2025-07-17T08:34:51.9685718Z * [new tag] trunk/134dfb3fe64323d7c615e866c43ea346ad987556 -> trunk/134dfb3fe64323d7c615e866c43ea346ad987556 2025-07-17T08:34:51.9687980Z * [new tag] trunk/1393f71e0714d067ed5ec1f57f135431d20619bc -> trunk/1393f71e0714d067ed5ec1f57f135431d20619bc 2025-07-17T08:34:51.9690320Z * [new tag] trunk/13bf2655c1404aff64e05e50dd7b0ce4c8dc0fcf -> trunk/13bf2655c1404aff64e05e50dd7b0ce4c8dc0fcf 2025-07-17T08:34:51.9692615Z * [new tag] trunk/13ea0f2c0a06101d9aab9ce12c30043be7c92d38 -> trunk/13ea0f2c0a06101d9aab9ce12c30043be7c92d38 2025-07-17T08:34:51.9694874Z * [new tag] trunk/13efb2c858c41141776ad42f7c44f1709948de71 -> trunk/13efb2c858c41141776ad42f7c44f1709948de71 2025-07-17T08:34:51.9697290Z * [new tag] trunk/144965ca9af478515736665b0577cded22fa692e -> trunk/144965ca9af478515736665b0577cded22fa692e 2025-07-17T08:34:51.9699599Z * [new tag] trunk/14ecc0336185f2ca5591858bc74cd4aadf2d1161 -> trunk/14ecc0336185f2ca5591858bc74cd4aadf2d1161 2025-07-17T08:34:51.9701874Z * [new tag] trunk/14f3639e09d692e53c9b9714776e3ca48fed7c4c -> trunk/14f3639e09d692e53c9b9714776e3ca48fed7c4c 2025-07-17T08:34:51.9704172Z * [new tag] trunk/154a39bfbd6e0cc6b7e6f3bce708ab0157ce1c50 -> trunk/154a39bfbd6e0cc6b7e6f3bce708ab0157ce1c50 2025-07-17T08:34:51.9706421Z * [new tag] trunk/156a377f4cf9b5b5255575e26d27f745c111a6ae -> trunk/156a377f4cf9b5b5255575e26d27f745c111a6ae 2025-07-17T08:34:51.9708840Z * [new tag] trunk/156b28e62a225214a5685088a848c5efe6a4b95e -> trunk/156b28e62a225214a5685088a848c5efe6a4b95e 2025-07-17T08:34:51.9711332Z * [new tag] trunk/156bc243f0edbf79d4a24d5affb48198551b58dd -> trunk/156bc243f0edbf79d4a24d5affb48198551b58dd 2025-07-17T08:34:51.9713409Z * [new tag] trunk/157683d8623a683a7e05755176f86a41c7bf518a -> trunk/157683d8623a683a7e05755176f86a41c7bf518a 2025-07-17T08:34:51.9716046Z * [new tag] trunk/1586521461c8dc642735466fc143b7d366a858d0 -> trunk/1586521461c8dc642735466fc143b7d366a858d0 2025-07-17T08:34:51.9718494Z * [new tag] trunk/159a39ad344cca84347e7b3be653ec13834cefdc -> trunk/159a39ad344cca84347e7b3be653ec13834cefdc 2025-07-17T08:34:51.9720718Z * [new tag] trunk/162ca185ff06fc7440e9f52d249e0b465976449a -> trunk/162ca185ff06fc7440e9f52d249e0b465976449a 2025-07-17T08:34:51.9723095Z * [new tag] trunk/163f0d8f2ab0a602a16f606db6d873298088e3a7 -> trunk/163f0d8f2ab0a602a16f606db6d873298088e3a7 2025-07-17T08:34:51.9725488Z * [new tag] trunk/16c3b4143b6fdabbee45ea5a64bb922a5969145a -> trunk/16c3b4143b6fdabbee45ea5a64bb922a5969145a 2025-07-17T08:34:51.9727786Z * [new tag] trunk/172853547ac842bb5ed8bf6e07702c76ff2cf043 -> trunk/172853547ac842bb5ed8bf6e07702c76ff2cf043 2025-07-17T08:34:51.9730006Z * [new tag] trunk/178fe7aa98987111a73534375099f4ad255e8b59 -> trunk/178fe7aa98987111a73534375099f4ad255e8b59 2025-07-17T08:34:51.9732320Z * [new tag] trunk/179dcc10e4e0c742fb7d93b832021d0c177798bf -> trunk/179dcc10e4e0c742fb7d93b832021d0c177798bf 2025-07-17T08:34:51.9734571Z * [new tag] trunk/17b38b850e5362c2f5d7cd75e6552e3d149232f2 -> trunk/17b38b850e5362c2f5d7cd75e6552e3d149232f2 2025-07-17T08:34:51.9736882Z * [new tag] trunk/17eb649d5596c52bae65a069e03b4550155ad57f -> trunk/17eb649d5596c52bae65a069e03b4550155ad57f 2025-07-17T08:34:51.9739133Z * [new tag] trunk/1839e8d04b81ee6eda0cff6fbfc218a7a600f6f7 -> trunk/1839e8d04b81ee6eda0cff6fbfc218a7a600f6f7 2025-07-17T08:34:51.9741464Z * [new tag] trunk/1851f50866a11e7ddede5d07c583e2c5407e7708 -> trunk/1851f50866a11e7ddede5d07c583e2c5407e7708 2025-07-17T08:34:51.9743713Z * [new tag] trunk/187828dcb4145cdfda916ff670207bfb17abdef8 -> trunk/187828dcb4145cdfda916ff670207bfb17abdef8 2025-07-17T08:34:51.9746079Z * [new tag] trunk/18b01afa9ed1dbc696a06de1b69bf5c021c18c10 -> trunk/18b01afa9ed1dbc696a06de1b69bf5c021c18c10 2025-07-17T08:34:51.9748294Z * [new tag] trunk/18bf6addc4741852953ac0b8d555411c347e67aa -> trunk/18bf6addc4741852953ac0b8d555411c347e67aa 2025-07-17T08:34:51.9750535Z * [new tag] trunk/18e4c461fb4e2bbdef889c3b1a2cdb29c0fb31cd -> trunk/18e4c461fb4e2bbdef889c3b1a2cdb29c0fb31cd 2025-07-17T08:34:51.9752828Z * [new tag] trunk/190f76fa313410df8dbb4111c586a516bf55515c -> trunk/190f76fa313410df8dbb4111c586a516bf55515c 2025-07-17T08:34:51.9756331Z * [new tag] trunk/1913c915e0d2b08e7af00889dc50e78d0a9b9198 -> trunk/1913c915e0d2b08e7af00889dc50e78d0a9b9198 2025-07-17T08:34:51.9758628Z * [new tag] trunk/191693ac854de1bb5fcf4280b7a93ab29cc12b5c -> trunk/191693ac854de1bb5fcf4280b7a93ab29cc12b5c 2025-07-17T08:34:51.9760870Z * [new tag] trunk/194539e9c33dc793fe67fbb68c7cee12f399e276 -> trunk/194539e9c33dc793fe67fbb68c7cee12f399e276 2025-07-17T08:34:51.9763158Z * [new tag] trunk/194c221e0abc9daf178e14ca2608c5d6804a3eb7 -> trunk/194c221e0abc9daf178e14ca2608c5d6804a3eb7 2025-07-17T08:34:51.9765559Z * [new tag] trunk/195ef1bce8f17419139fb81406b91f476eba7257 -> trunk/195ef1bce8f17419139fb81406b91f476eba7257 2025-07-17T08:34:51.9767796Z * [new tag] trunk/19625daf889f0a6192a76e200205817e3ee27f26 -> trunk/19625daf889f0a6192a76e200205817e3ee27f26 2025-07-17T08:34:51.9770219Z * [new tag] trunk/196c95d463367f15999c0cddc9eb89031e9988ab -> trunk/196c95d463367f15999c0cddc9eb89031e9988ab 2025-07-17T08:34:51.9772581Z * [new tag] trunk/197c1869f5467b4cfa1e197bbe394be82ba40554 -> trunk/197c1869f5467b4cfa1e197bbe394be82ba40554 2025-07-17T08:34:51.9774995Z * [new tag] trunk/1982ec2d22c5145e4ffeec08064c5ca17e969c25 -> trunk/1982ec2d22c5145e4ffeec08064c5ca17e969c25 2025-07-17T08:34:51.9777107Z * [new tag] trunk/19a01382bc623cf30cdfa7215e47a6b69d8598ff -> trunk/19a01382bc623cf30cdfa7215e47a6b69d8598ff 2025-07-17T08:34:51.9779238Z * [new tag] trunk/19ae5afdaa3a59135e6adbaea197b7a2456f1865 -> trunk/19ae5afdaa3a59135e6adbaea197b7a2456f1865 2025-07-17T08:34:51.9781372Z * [new tag] trunk/19f851ce10b16f0ed11d18d937ca7b32746153b0 -> trunk/19f851ce10b16f0ed11d18d937ca7b32746153b0 2025-07-17T08:34:51.9783737Z * [new tag] trunk/19ffb5e6f7606436249742b0f3efc0bab244dc55 -> trunk/19ffb5e6f7606436249742b0f3efc0bab244dc55 2025-07-17T08:34:51.9786280Z * [new tag] trunk/19ffdf4ea053bc5befe7eaa737fd1fafaaeb34a2 -> trunk/19ffdf4ea053bc5befe7eaa737fd1fafaaeb34a2 2025-07-17T08:34:51.9788577Z * [new tag] trunk/1a195bf7d6155b027f7ca81a97d71a3af34ce11e -> trunk/1a195bf7d6155b027f7ca81a97d71a3af34ce11e 2025-07-17T08:34:51.9790849Z * [new tag] trunk/1a568f4e5d1c0e2e8994d64f01e1e8a3104a1a48 -> trunk/1a568f4e5d1c0e2e8994d64f01e1e8a3104a1a48 2025-07-17T08:34:51.9793070Z * [new tag] trunk/1b032384b186154bb140c7368e2e700cb1419b86 -> trunk/1b032384b186154bb140c7368e2e700cb1419b86 2025-07-17T08:34:51.9795613Z * [new tag] trunk/1b389025ba0cc640e07991314bfba8b6ca385bd2 -> trunk/1b389025ba0cc640e07991314bfba8b6ca385bd2 2025-07-17T08:34:51.9797885Z * [new tag] trunk/1b3d69b59f92383633731aada8383ab88da3ed60 -> trunk/1b3d69b59f92383633731aada8383ab88da3ed60 2025-07-17T08:34:51.9800215Z * [new tag] trunk/1b50c12584909bda00009f4f0fd0d38ec792d019 -> trunk/1b50c12584909bda00009f4f0fd0d38ec792d019 2025-07-17T08:34:51.9802618Z * [new tag] trunk/1b58e7adab91fe20bbfb1568403d72869317e75c -> trunk/1b58e7adab91fe20bbfb1568403d72869317e75c 2025-07-17T08:34:51.9804927Z * [new tag] trunk/1b6772a90f5c334cab7b9d055a7f819ce4c3478c -> trunk/1b6772a90f5c334cab7b9d055a7f819ce4c3478c 2025-07-17T08:34:51.9807264Z * [new tag] trunk/1bb9b1858b84ee07ff1bc33157d4ea7a980bb6b5 -> trunk/1bb9b1858b84ee07ff1bc33157d4ea7a980bb6b5 2025-07-17T08:34:51.9809506Z * [new tag] trunk/1c6057fd179b0373686a790b0a0b7fc68fe7f27d -> trunk/1c6057fd179b0373686a790b0a0b7fc68fe7f27d 2025-07-17T08:34:51.9811800Z * [new tag] trunk/1c8844d9e7b2d72fb80b67ed51df4f6a1295b3b5 -> trunk/1c8844d9e7b2d72fb80b67ed51df4f6a1295b3b5 2025-07-17T08:34:51.9814085Z * [new tag] trunk/1c960c5638b6f20f7a1cbe1e9036c14a4845a195 -> trunk/1c960c5638b6f20f7a1cbe1e9036c14a4845a195 2025-07-17T08:34:51.9816337Z * [new tag] trunk/1cb0597a890bc50b8ed3937d6bdf5a734691bd91 -> trunk/1cb0597a890bc50b8ed3937d6bdf5a734691bd91 2025-07-17T08:34:51.9818471Z * [new tag] trunk/1cc62c2cb91e56ae50494f88f369cd6ec466a118 -> trunk/1cc62c2cb91e56ae50494f88f369cd6ec466a118 2025-07-17T08:34:51.9820827Z * [new tag] trunk/1ccc57e4288e231210c9b2c29943b2752259bf44 -> trunk/1ccc57e4288e231210c9b2c29943b2752259bf44 2025-07-17T08:34:51.9823098Z * [new tag] trunk/1cce73b5f4806f266fbbcf3383057af5f2e3a0aa -> trunk/1cce73b5f4806f266fbbcf3383057af5f2e3a0aa 2025-07-17T08:34:51.9825370Z * [new tag] trunk/1cfdcb975a3b0c685c34c6ae0e378b9bc92e0050 -> trunk/1cfdcb975a3b0c685c34c6ae0e378b9bc92e0050 2025-07-17T08:34:51.9827572Z * [new tag] trunk/1d0f45d5d15389b83312b6942ac2017c8d7ebedc -> trunk/1d0f45d5d15389b83312b6942ac2017c8d7ebedc 2025-07-17T08:34:51.9829941Z * [new tag] trunk/1d584761622ff6e5519c5e3dbbb62a21b89ffe8a -> trunk/1d584761622ff6e5519c5e3dbbb62a21b89ffe8a 2025-07-17T08:34:51.9832187Z * [new tag] trunk/1d993fa3092e4f0b5745f2470024b35cac96da14 -> trunk/1d993fa3092e4f0b5745f2470024b35cac96da14 2025-07-17T08:34:51.9834753Z * [new tag] trunk/1dc1eedd4369f6e6bb79d5315e3ffc1bdc59b709 -> trunk/1dc1eedd4369f6e6bb79d5315e3ffc1bdc59b709 2025-07-17T08:34:51.9836912Z * [new tag] trunk/1dd0b1d12ba48d7879a57391cab6213742dcadb6 -> trunk/1dd0b1d12ba48d7879a57391cab6213742dcadb6 2025-07-17T08:34:51.9839231Z * [new tag] trunk/1e373d02d57ff65c70febc4ec427bf6d1708483e -> trunk/1e373d02d57ff65c70febc4ec427bf6d1708483e 2025-07-17T08:34:51.9841540Z * [new tag] trunk/1e474cc9c8ddab22ed314ad8be641002b0161498 -> trunk/1e474cc9c8ddab22ed314ad8be641002b0161498 2025-07-17T08:34:51.9843857Z * [new tag] trunk/1e4c5b666afe6434cbe3b830aaf70864074d40f5 -> trunk/1e4c5b666afe6434cbe3b830aaf70864074d40f5 2025-07-17T08:34:51.9846231Z * [new tag] trunk/1e4d8b5a4a67220473bf0027c58baaa08a036714 -> trunk/1e4d8b5a4a67220473bf0027c58baaa08a036714 2025-07-17T08:34:51.9848507Z * [new tag] trunk/1e690b6c41677bf1fe3147e3fc011eb9805365f5 -> trunk/1e690b6c41677bf1fe3147e3fc011eb9805365f5 2025-07-17T08:34:51.9850786Z * [new tag] trunk/1e6a653234c299ccecf702233a1eeb7455b1ce77 -> trunk/1e6a653234c299ccecf702233a1eeb7455b1ce77 2025-07-17T08:34:51.9852886Z * [new tag] trunk/1e7e21ec5dd6ea28b27c57fe92bcf31ea5983913 -> trunk/1e7e21ec5dd6ea28b27c57fe92bcf31ea5983913 2025-07-17T08:34:51.9854982Z * [new tag] trunk/1e8e9f745e43fa38bbfc7b67b30bc66c0e7ebbd6 -> trunk/1e8e9f745e43fa38bbfc7b67b30bc66c0e7ebbd6 2025-07-17T08:34:51.9857332Z * [new tag] trunk/1ea9cde598ead20194dbb6c5cb26e74e36e6ad55 -> trunk/1ea9cde598ead20194dbb6c5cb26e74e36e6ad55 2025-07-17T08:34:51.9859569Z * [new tag] trunk/1ed243f01c8efb329055c6124ba0aa5f48747cfe -> trunk/1ed243f01c8efb329055c6124ba0aa5f48747cfe 2025-07-17T08:34:51.9861912Z * [new tag] trunk/1eea2c4fe35ffbdcbfccbeb7ac6c3ec02137385d -> trunk/1eea2c4fe35ffbdcbfccbeb7ac6c3ec02137385d 2025-07-17T08:34:51.9864117Z * [new tag] trunk/1f0eb79e3e91f09f99b53aa4d331eb47a1f66101 -> trunk/1f0eb79e3e91f09f99b53aa4d331eb47a1f66101 2025-07-17T08:34:51.9866477Z * [new tag] trunk/1f1f22991dcca01ed27fad44c21f0827d47bdd6b -> trunk/1f1f22991dcca01ed27fad44c21f0827d47bdd6b 2025-07-17T08:34:51.9868770Z * [new tag] trunk/1f3cc4875cb9d9b579a41b6699f97820c16c3868 -> trunk/1f3cc4875cb9d9b579a41b6699f97820c16c3868 2025-07-17T08:34:51.9870878Z * [new tag] trunk/1f57e0e04da9d334e238cec346f7ae3667bed9d1 -> trunk/1f57e0e04da9d334e238cec346f7ae3667bed9d1 2025-07-17T08:34:51.9873270Z * [new tag] trunk/1fff6356d9f99ec2980052c498dc0b70d513bc6d -> trunk/1fff6356d9f99ec2980052c498dc0b70d513bc6d 2025-07-17T08:34:51.9875855Z * [new tag] trunk/2002e3a311968d18afd0fa5d195943676f134fa2 -> trunk/2002e3a311968d18afd0fa5d195943676f134fa2 2025-07-17T08:34:51.9878157Z * [new tag] trunk/2022588295295fa1fdee7a500adad3bd0c28b9d6 -> trunk/2022588295295fa1fdee7a500adad3bd0c28b9d6 2025-07-17T08:34:51.9880407Z * [new tag] trunk/202d2ae53a493700da813ed7d6055c0b62c7fc6c -> trunk/202d2ae53a493700da813ed7d6055c0b62c7fc6c 2025-07-17T08:34:51.9882727Z * [new tag] trunk/204db27a0c4478c36dbb15fa2d4caed783a90265 -> trunk/204db27a0c4478c36dbb15fa2d4caed783a90265 2025-07-17T08:34:51.9884940Z * [new tag] trunk/205241a0d5149d05e44dc113dc0273e8eceff9f0 -> trunk/205241a0d5149d05e44dc113dc0273e8eceff9f0 2025-07-17T08:34:51.9887263Z * [new tag] trunk/208ec60e72a63f366c757a5bc895089ceb323fcc -> trunk/208ec60e72a63f366c757a5bc895089ceb323fcc 2025-07-17T08:34:51.9889628Z * [new tag] trunk/20a74c370b8427cce369b1abf0c470344d01dc8e -> trunk/20a74c370b8427cce369b1abf0c470344d01dc8e 2025-07-17T08:34:51.9891892Z * [new tag] trunk/20e40492b046b9287726d3ec656117e4dc38f0e2 -> trunk/20e40492b046b9287726d3ec656117e4dc38f0e2 2025-07-17T08:34:51.9894243Z * [new tag] trunk/210632fae1aa2cd994645a70836ff6c602e1beb9 -> trunk/210632fae1aa2cd994645a70836ff6c602e1beb9 2025-07-17T08:34:51.9896384Z * [new tag] trunk/212575f994abbc362da2517510c49f7a34e0a838 -> trunk/212575f994abbc362da2517510c49f7a34e0a838 2025-07-17T08:34:51.9898621Z * [new tag] trunk/214e2959dcdbf91a999d5c0a5d40c91e4442e8c5 -> trunk/214e2959dcdbf91a999d5c0a5d40c91e4442e8c5 2025-07-17T08:34:51.9900853Z * [new tag] trunk/2161be849702330509b452fefc2e103d3be54cd4 -> trunk/2161be849702330509b452fefc2e103d3be54cd4 2025-07-17T08:34:51.9903108Z * [new tag] trunk/216bd6091ec52865052282eced7e6d5d2a4b4fb4 -> trunk/216bd6091ec52865052282eced7e6d5d2a4b4fb4 2025-07-17T08:34:51.9905343Z * [new tag] trunk/2179afd7149c117dace9e552419082094b10a386 -> trunk/2179afd7149c117dace9e552419082094b10a386 2025-07-17T08:34:51.9907583Z * [new tag] trunk/21990fbad97acec769f737b450033774c7be8737 -> trunk/21990fbad97acec769f737b450033774c7be8737 2025-07-17T08:34:51.9909865Z * [new tag] trunk/21b5dc7a6aadb5da6c3f5be61f6204719dae00ca -> trunk/21b5dc7a6aadb5da6c3f5be61f6204719dae00ca 2025-07-17T08:34:51.9912180Z * [new tag] trunk/22edb457c9bf13e0d66631dff4f4ace5c38e0f3b -> trunk/22edb457c9bf13e0d66631dff4f4ace5c38e0f3b 2025-07-17T08:34:51.9914946Z * [new tag] trunk/22f3347fd9482bdf1006c9aa92353b739b7e868b -> trunk/22f3347fd9482bdf1006c9aa92353b739b7e868b 2025-07-17T08:34:51.9917341Z * [new tag] trunk/231eb9902ba78a4ef70203243058f3c7c0ced15d -> trunk/231eb9902ba78a4ef70203243058f3c7c0ced15d 2025-07-17T08:34:51.9919699Z * [new tag] trunk/23491519d288dedb2a54cfad5fef7fcb2ad8eade -> trunk/23491519d288dedb2a54cfad5fef7fcb2ad8eade 2025-07-17T08:34:51.9921938Z * [new tag] trunk/2380115f9738f97cf706affefd647d2cb6dfbb3f -> trunk/2380115f9738f97cf706affefd647d2cb6dfbb3f 2025-07-17T08:34:51.9924290Z * [new tag] trunk/24063ad10994e728087b3958cae9cc27dd376630 -> trunk/24063ad10994e728087b3958cae9cc27dd376630 2025-07-17T08:34:51.9926548Z * [new tag] trunk/242eb19c8383b4b197963a8a564475d52c85ac66 -> trunk/242eb19c8383b4b197963a8a564475d52c85ac66 2025-07-17T08:34:51.9929017Z * [new tag] trunk/243b12e5657a516d6e7b1a0a3f55851ce99bd4cb -> trunk/243b12e5657a516d6e7b1a0a3f55851ce99bd4cb 2025-07-17T08:34:51.9931367Z * [new tag] trunk/247113e03e3fc3e933138b52f856f33a8be33071 -> trunk/247113e03e3fc3e933138b52f856f33a8be33071 2025-07-17T08:34:51.9933682Z * [new tag] trunk/2471cc33550fc00164e2759b1d9928960e9cfaf6 -> trunk/2471cc33550fc00164e2759b1d9928960e9cfaf6 2025-07-17T08:34:51.9936005Z * [new tag] trunk/247f83e0a475bc7eea97c76719f8b5ffa87edced -> trunk/247f83e0a475bc7eea97c76719f8b5ffa87edced 2025-07-17T08:34:51.9938302Z * [new tag] trunk/2481c4b2ea5db7a6d6f60f5d27426000c2779dc8 -> trunk/2481c4b2ea5db7a6d6f60f5d27426000c2779dc8 2025-07-17T08:34:51.9940568Z * [new tag] trunk/24b49b98810bb77f3cfa4c15baa9a15c9be3db61 -> trunk/24b49b98810bb77f3cfa4c15baa9a15c9be3db61 2025-07-17T08:34:51.9942798Z * [new tag] trunk/250ae2531c55dcc50f558ec739941324e3f9a4d4 -> trunk/250ae2531c55dcc50f558ec739941324e3f9a4d4 2025-07-17T08:34:51.9945055Z * [new tag] trunk/25717da8c869a06befed60bc1a0e3dcdac7fd7d9 -> trunk/25717da8c869a06befed60bc1a0e3dcdac7fd7d9 2025-07-17T08:34:51.9947297Z * [new tag] trunk/2578796e235d1d878272084253bee3e2cc02a5b1 -> trunk/2578796e235d1d878272084253bee3e2cc02a5b1 2025-07-17T08:34:51.9949593Z * [new tag] trunk/2585960b47d6429a119665216123cb00435efde9 -> trunk/2585960b47d6429a119665216123cb00435efde9 2025-07-17T08:34:51.9951852Z * [new tag] trunk/2596e3d0617852469241be8777cf46db5c83928c -> trunk/2596e3d0617852469241be8777cf46db5c83928c 2025-07-17T08:34:51.9954239Z * [new tag] trunk/2620361d19f9c4bf37a71c8477823d605191c93a -> trunk/2620361d19f9c4bf37a71c8477823d605191c93a 2025-07-17T08:34:51.9957407Z * [new tag] trunk/2625c70aecc6eced1dbe108279feab7509733bef -> trunk/2625c70aecc6eced1dbe108279feab7509733bef 2025-07-17T08:34:51.9959505Z * [new tag] trunk/262654ee518eb314678f53baf4e133e4767eca3d -> trunk/262654ee518eb314678f53baf4e133e4767eca3d 2025-07-17T08:34:51.9961756Z * [new tag] trunk/26807dcf277feb2d99ab88d7b6da526488baea93 -> trunk/26807dcf277feb2d99ab88d7b6da526488baea93 2025-07-17T08:34:51.9964020Z * [new tag] trunk/26f066bb614c2ee5c0eff598c97855b1df011ef8 -> trunk/26f066bb614c2ee5c0eff598c97855b1df011ef8 2025-07-17T08:34:51.9966385Z * [new tag] trunk/26f7ca39724ff43b7b1e22190feeea67d2212f2d -> trunk/26f7ca39724ff43b7b1e22190feeea67d2212f2d 2025-07-17T08:34:51.9968705Z * [new tag] trunk/271ca679a8a0e40f1ecca49db76b5f8d6a9713d6 -> trunk/271ca679a8a0e40f1ecca49db76b5f8d6a9713d6 2025-07-17T08:34:51.9970830Z * [new tag] trunk/276c790010b015b79a7ca110e28f8e2111cfdb79 -> trunk/276c790010b015b79a7ca110e28f8e2111cfdb79 2025-07-17T08:34:51.9973059Z * [new tag] trunk/2796f31b5e3c90268365e961e2374df3ea93ff53 -> trunk/2796f31b5e3c90268365e961e2374df3ea93ff53 2025-07-17T08:34:51.9975336Z * [new tag] trunk/279cae52e750d5b4e0b49f3f1abca6c2027dd063 -> trunk/279cae52e750d5b4e0b49f3f1abca6c2027dd063 2025-07-17T08:34:51.9977667Z * [new tag] trunk/27c50799c1ca156166bf49ce72a5e97097a44d7d -> trunk/27c50799c1ca156166bf49ce72a5e97097a44d7d 2025-07-17T08:34:51.9979996Z * [new tag] trunk/27df0c56b7c6b75a5f77f5714aeef2ef8f1faa2c -> trunk/27df0c56b7c6b75a5f77f5714aeef2ef8f1faa2c 2025-07-17T08:34:51.9982237Z * [new tag] trunk/2815ade9a80e1f557370e86500b21cbc465a8ffe -> trunk/2815ade9a80e1f557370e86500b21cbc465a8ffe 2025-07-17T08:34:51.9984522Z * [new tag] trunk/2815eea0d0e02eaa0243fa8d1b9d8935b13b3aa5 -> trunk/2815eea0d0e02eaa0243fa8d1b9d8935b13b3aa5 2025-07-17T08:34:51.9986901Z * [new tag] trunk/2860f5c4f5328224ab7998bc90b9fb395b5b068b -> trunk/2860f5c4f5328224ab7998bc90b9fb395b5b068b 2025-07-17T08:34:51.9989208Z * [new tag] trunk/28796f71d04302029290f473a286efc2aba339c2 -> trunk/28796f71d04302029290f473a286efc2aba339c2 2025-07-17T08:34:51.9991343Z * [new tag] trunk/28aae93f24556c6aa8c972c0479b01b908b67fbe -> trunk/28aae93f24556c6aa8c972c0479b01b908b67fbe 2025-07-17T08:34:51.9993705Z * [new tag] trunk/2903e5ad3c4304b90daa6e5ed44a379225571b3a -> trunk/2903e5ad3c4304b90daa6e5ed44a379225571b3a 2025-07-17T08:34:51.9996108Z * [new tag] trunk/2908c10259bac21b00e9b36318e364801e0ae910 -> trunk/2908c10259bac21b00e9b36318e364801e0ae910 2025-07-17T08:34:51.9998433Z * [new tag] trunk/29391c7cf927857123108fa5e9dae6ed3db489f1 -> trunk/29391c7cf927857123108fa5e9dae6ed3db489f1 2025-07-17T08:34:52.0000684Z * [new tag] trunk/297805fd8f59b76a28048a79e8bced2616ed8713 -> trunk/297805fd8f59b76a28048a79e8bced2616ed8713 2025-07-17T08:34:52.0002987Z * [new tag] trunk/297daa1d30c80826b939d8f2dcd07422dec72642 -> trunk/297daa1d30c80826b939d8f2dcd07422dec72642 2025-07-17T08:34:52.0005324Z * [new tag] trunk/29867b211ab74355a0e03bdc9e204d89668c291a -> trunk/29867b211ab74355a0e03bdc9e204d89668c291a 2025-07-17T08:34:52.0007712Z * [new tag] trunk/29e6033ff38f0deb8dde1146a1612e64ed00e3b7 -> trunk/29e6033ff38f0deb8dde1146a1612e64ed00e3b7 2025-07-17T08:34:52.0009962Z * [new tag] trunk/29f76ec0f3eccf619a0aee03e3abbd4914a1b4b2 -> trunk/29f76ec0f3eccf619a0aee03e3abbd4914a1b4b2 2025-07-17T08:34:52.0012200Z * [new tag] trunk/2a3b41cbd077b229bbdd59f6f068c9b643a78b0b -> trunk/2a3b41cbd077b229bbdd59f6f068c9b643a78b0b 2025-07-17T08:34:52.0014481Z * [new tag] trunk/2a4e357192a2b01c9acaeb2168349e5396c9192a -> trunk/2a4e357192a2b01c9acaeb2168349e5396c9192a 2025-07-17T08:34:52.0016933Z * [new tag] trunk/2a8795a981c02f57a57f8699d0e85e42c53b6117 -> trunk/2a8795a981c02f57a57f8699d0e85e42c53b6117 2025-07-17T08:34:52.0019062Z * [new tag] trunk/2aade5ee9fafe52aef4e60624e6c950bc25f8a3e -> trunk/2aade5ee9fafe52aef4e60624e6c950bc25f8a3e 2025-07-17T08:34:52.0021126Z * [new tag] trunk/2ad5c25cfc603c3656e6699d6137419dbb009495 -> trunk/2ad5c25cfc603c3656e6699d6137419dbb009495 2025-07-17T08:34:52.0023461Z * [new tag] trunk/2af7c67e48cad59d648a3e85501f74eade5a3268 -> trunk/2af7c67e48cad59d648a3e85501f74eade5a3268 2025-07-17T08:34:52.0025712Z * [new tag] trunk/2b0f9b1f6172a0d5817c7ac7406200897311da5f -> trunk/2b0f9b1f6172a0d5817c7ac7406200897311da5f 2025-07-17T08:34:52.0028054Z * [new tag] trunk/2b19d85d70ecf1b7121e6d35f2adff06e1eb2e06 -> trunk/2b19d85d70ecf1b7121e6d35f2adff06e1eb2e06 2025-07-17T08:34:52.0030339Z * [new tag] trunk/2b9d638e3333e6e9ae324e1486774e83292e1883 -> trunk/2b9d638e3333e6e9ae324e1486774e83292e1883 2025-07-17T08:34:52.0032561Z * [new tag] trunk/2ba930d4ce04f417195f3061e55479a1a8b16b9d -> trunk/2ba930d4ce04f417195f3061e55479a1a8b16b9d 2025-07-17T08:34:52.0035012Z * [new tag] trunk/2bb33e7a08c4710024b3cb249a13cd2f5d0f6473 -> trunk/2bb33e7a08c4710024b3cb249a13cd2f5d0f6473 2025-07-17T08:34:52.0037280Z * [new tag] trunk/2c0d94a7debe83e96a821548a6e0cb18a0f5cfc3 -> trunk/2c0d94a7debe83e96a821548a6e0cb18a0f5cfc3 2025-07-17T08:34:52.0039610Z * [new tag] trunk/2c1a93a0ae7ebd7963a5931d7b34c885d57844d9 -> trunk/2c1a93a0ae7ebd7963a5931d7b34c885d57844d9 2025-07-17T08:34:52.0041930Z * [new tag] trunk/2c6324a1ebf3e3f6a12be6d18fc0bcdbe00f2a0a -> trunk/2c6324a1ebf3e3f6a12be6d18fc0bcdbe00f2a0a 2025-07-17T08:34:52.0044233Z * [new tag] trunk/2c68c3e8d5e9a235f5861be6486de4959f80c840 -> trunk/2c68c3e8d5e9a235f5861be6486de4959f80c840 2025-07-17T08:34:52.0046550Z * [new tag] trunk/2c76f31221e117b217b8a6a96a5405f626d2218a -> trunk/2c76f31221e117b217b8a6a96a5405f626d2218a 2025-07-17T08:34:52.0048806Z * [new tag] trunk/2cdcd16e836a144f36996104b1340b03041eb07a -> trunk/2cdcd16e836a144f36996104b1340b03041eb07a 2025-07-17T08:34:52.0051053Z * [new tag] trunk/2d3615f577894c7a117a55e85bb8371bb598ec50 -> trunk/2d3615f577894c7a117a55e85bb8371bb598ec50 2025-07-17T08:34:52.0053155Z * [new tag] trunk/2d39a48d524021995269411bd49fe792e59d9f94 -> trunk/2d39a48d524021995269411bd49fe792e59d9f94 2025-07-17T08:34:52.0055539Z * [new tag] trunk/2d7e6c6241971106a56073d7a53c7d1336b11a51 -> trunk/2d7e6c6241971106a56073d7a53c7d1336b11a51 2025-07-17T08:34:52.0057835Z * [new tag] trunk/2d832c9587fd99db295b62d0c9b459d509c19d06 -> trunk/2d832c9587fd99db295b62d0c9b459d509c19d06 2025-07-17T08:34:52.0060199Z * [new tag] trunk/2db1a54465b6121993ad7827f1327ed319c81d70 -> trunk/2db1a54465b6121993ad7827f1327ed319c81d70 2025-07-17T08:34:52.0062503Z * [new tag] trunk/2dc16274519841dc2778cba6962cb0357d1ee8dc -> trunk/2dc16274519841dc2778cba6962cb0357d1ee8dc 2025-07-17T08:34:52.0064736Z * [new tag] trunk/2e0e08588e7cb04f49ee5bbc000b18b72864dfa1 -> trunk/2e0e08588e7cb04f49ee5bbc000b18b72864dfa1 2025-07-17T08:34:52.0067134Z * [new tag] trunk/2e14069081172faf9a51bd34c1e0b85dbf39cb4b -> trunk/2e14069081172faf9a51bd34c1e0b85dbf39cb4b 2025-07-17T08:34:52.0069426Z * [new tag] trunk/2e2ea7290a1cf2da3c3efd6e6ad4836bb94beaeb -> trunk/2e2ea7290a1cf2da3c3efd6e6ad4836bb94beaeb 2025-07-17T08:34:52.0071666Z * [new tag] trunk/2e64e45b0b260e2efc2bde641ec2b3b8ed910194 -> trunk/2e64e45b0b260e2efc2bde641ec2b3b8ed910194 2025-07-17T08:34:52.0073963Z * [new tag] trunk/2e9bd03f6075b9b93b070c3770bf58f60be5508e -> trunk/2e9bd03f6075b9b93b070c3770bf58f60be5508e 2025-07-17T08:34:52.0076322Z * [new tag] trunk/2eb744c08d600e84b167dbda7daa792243a2c235 -> trunk/2eb744c08d600e84b167dbda7daa792243a2c235 2025-07-17T08:34:52.0078617Z * [new tag] trunk/2ee23175d96aa022a9c94209c14bbb7f8e064760 -> trunk/2ee23175d96aa022a9c94209c14bbb7f8e064760 2025-07-17T08:34:52.0080829Z * [new tag] trunk/2efa5eaa652f9a6505635d96178e3f119b20adf7 -> trunk/2efa5eaa652f9a6505635d96178e3f119b20adf7 2025-07-17T08:34:52.0083041Z * [new tag] trunk/2eff14c4455de22628ea599a0d4ba047a7eba6ae -> trunk/2eff14c4455de22628ea599a0d4ba047a7eba6ae 2025-07-17T08:34:52.0085356Z * [new tag] trunk/2f1c5c4131047d25f822ac278bac8c13deff4079 -> trunk/2f1c5c4131047d25f822ac278bac8c13deff4079 2025-07-17T08:34:52.0087807Z * [new tag] trunk/2f94f69b7c83370ef0cc65e3ab96bb5bf11a7b1a -> trunk/2f94f69b7c83370ef0cc65e3ab96bb5bf11a7b1a 2025-07-17T08:34:52.0090057Z * [new tag] trunk/2fde2090d001a9ae63ef41413f11897761eb4de9 -> trunk/2fde2090d001a9ae63ef41413f11897761eb4de9 2025-07-17T08:34:52.0092199Z * [new tag] trunk/2ff3280c77c705e11c5211d4be8fef9853cd0559 -> trunk/2ff3280c77c705e11c5211d4be8fef9853cd0559 2025-07-17T08:34:52.0094461Z * [new tag] trunk/3003c681ef1f8efbfa83470173c62930d195364b -> trunk/3003c681ef1f8efbfa83470173c62930d195364b 2025-07-17T08:34:52.0096833Z * [new tag] trunk/30293b8b5edca86cfc80f24874b0968620a8e72d -> trunk/30293b8b5edca86cfc80f24874b0968620a8e72d 2025-07-17T08:34:52.0099218Z * [new tag] trunk/30387ab2e485384ab2e67084a1e2c5569190ba92 -> trunk/30387ab2e485384ab2e67084a1e2c5569190ba92 2025-07-17T08:34:52.0101485Z * [new tag] trunk/3040ca6d0f8558e39919b14eebeacc34ddf980f5 -> trunk/3040ca6d0f8558e39919b14eebeacc34ddf980f5 2025-07-17T08:34:52.0103743Z * [new tag] trunk/30587195d314eb5eb02ce63f39a9be4c943629ef -> trunk/30587195d314eb5eb02ce63f39a9be4c943629ef 2025-07-17T08:34:52.0105992Z * [new tag] trunk/306dd19216b656467143483395ef582feb5d7d07 -> trunk/306dd19216b656467143483395ef582feb5d7d07 2025-07-17T08:34:52.0108331Z * [new tag] trunk/30a1cc11a47e63c4612ac862f1f2d984ab4dbf24 -> trunk/30a1cc11a47e63c4612ac862f1f2d984ab4dbf24 2025-07-17T08:34:52.0110614Z * [new tag] trunk/30d2648a4ae1585e70189e6fdca6dc9ac3d074ec -> trunk/30d2648a4ae1585e70189e6fdca6dc9ac3d074ec 2025-07-17T08:34:52.0112868Z * [new tag] trunk/30d3cf62fb813819364007e26b3eda84f7cddf3b -> trunk/30d3cf62fb813819364007e26b3eda84f7cddf3b 2025-07-17T08:34:52.0115593Z * [new tag] trunk/3106a33e410db9d1c2b0e47f9e329611d344a827 -> trunk/3106a33e410db9d1c2b0e47f9e329611d344a827 2025-07-17T08:34:52.0117861Z * [new tag] trunk/310e8361c565ca1602e719e4c812dc3931ec84d7 -> trunk/310e8361c565ca1602e719e4c812dc3931ec84d7 2025-07-17T08:34:52.0120162Z * [new tag] trunk/31326a9ad7db53e5bf4e03f8e52b75da944d0298 -> trunk/31326a9ad7db53e5bf4e03f8e52b75da944d0298 2025-07-17T08:34:52.0122442Z * [new tag] trunk/313a6a8ef94d689331b2bd8161f95c23d42eb22d -> trunk/313a6a8ef94d689331b2bd8161f95c23d42eb22d 2025-07-17T08:34:52.0124728Z * [new tag] trunk/3159ee2ad3e586573edb977c4c1745a26d9f1007 -> trunk/3159ee2ad3e586573edb977c4c1745a26d9f1007 2025-07-17T08:34:52.0126978Z * [new tag] trunk/31659964a5b546a84a4c18bb9a0828a0f63b8ccb -> trunk/31659964a5b546a84a4c18bb9a0828a0f63b8ccb 2025-07-17T08:34:52.0129251Z * [new tag] trunk/3173616532ecf0d5c0b78595d6054f1b4bd2bd75 -> trunk/3173616532ecf0d5c0b78595d6054f1b4bd2bd75 2025-07-17T08:34:52.0131511Z * [new tag] trunk/317520bf6ed5c1f2ef324f841e8ec0d8d80cb58a -> trunk/317520bf6ed5c1f2ef324f841e8ec0d8d80cb58a 2025-07-17T08:34:52.0133842Z * [new tag] trunk/317af4c87b7b8b03b8a08a4ead84d4065dd920e0 -> trunk/317af4c87b7b8b03b8a08a4ead84d4065dd920e0 2025-07-17T08:34:52.0135953Z * [new tag] trunk/31e127459789f9d97b3b35001c50e76f1b580cd5 -> trunk/31e127459789f9d97b3b35001c50e76f1b580cd5 2025-07-17T08:34:52.0138414Z * [new tag] trunk/3232b57cd87fbd15c990fbf87d181716a1993a55 -> trunk/3232b57cd87fbd15c990fbf87d181716a1993a55 2025-07-17T08:34:52.0140651Z * [new tag] trunk/326e751d07b92f788bece086227da9288206a27a -> trunk/326e751d07b92f788bece086227da9288206a27a 2025-07-17T08:34:52.0142912Z * [new tag] trunk/32983ea698842a2ca331eabb06acce9c99acb082 -> trunk/32983ea698842a2ca331eabb06acce9c99acb082 2025-07-17T08:34:52.0145281Z * [new tag] trunk/32c1611263ecc8832a8d7494480654ad851911c0 -> trunk/32c1611263ecc8832a8d7494480654ad851911c0 2025-07-17T08:34:52.0147565Z * [new tag] trunk/32eee8ed225d9f10fbbcb38c24b8b44c24c0c97c -> trunk/32eee8ed225d9f10fbbcb38c24b8b44c24c0c97c 2025-07-17T08:34:52.0150226Z * [new tag] trunk/3321acc92e24859dbe2ac6499067d1afde5622c3 -> trunk/3321acc92e24859dbe2ac6499067d1afde5622c3 2025-07-17T08:34:52.0152594Z * [new tag] trunk/333e0e61472c71cb2ad51a59cc5c56f6a47bc747 -> trunk/333e0e61472c71cb2ad51a59cc5c56f6a47bc747 2025-07-17T08:34:52.0155140Z * [new tag] trunk/3341c131b767a4036c152624c1e43baaf24cadf9 -> trunk/3341c131b767a4036c152624c1e43baaf24cadf9 2025-07-17T08:34:52.0157496Z * [new tag] trunk/336bff6d58ceb50b12d9d67764fd9f238bc0adb5 -> trunk/336bff6d58ceb50b12d9d67764fd9f238bc0adb5 2025-07-17T08:34:52.0159762Z * [new tag] trunk/336f1e2d35dfc63925754e884543491cc3f98f8c -> trunk/336f1e2d35dfc63925754e884543491cc3f98f8c 2025-07-17T08:34:52.0162046Z * [new tag] trunk/338a8c7853ab9b3e67dbcb402d30fec066bdd856 -> trunk/338a8c7853ab9b3e67dbcb402d30fec066bdd856 2025-07-17T08:34:52.0164348Z * [new tag] trunk/3404c1f0cfe4c8fd41d718d38316c24a3c0474b0 -> trunk/3404c1f0cfe4c8fd41d718d38316c24a3c0474b0 2025-07-17T08:34:52.0166579Z * [new tag] trunk/344731fb257440ab04f2d634929821be9be3fa80 -> trunk/344731fb257440ab04f2d634929821be9be3fa80 2025-07-17T08:34:52.0169704Z * [new tag] trunk/347ace4c7ac2dbb14799089c30bd01a9ac312791 -> trunk/347ace4c7ac2dbb14799089c30bd01a9ac312791 2025-07-17T08:34:52.0171977Z * [new tag] trunk/348e2a76dfe9c8a1995ab6ce4902dbeba935c984 -> trunk/348e2a76dfe9c8a1995ab6ce4902dbeba935c984 2025-07-17T08:34:52.0174259Z * [new tag] trunk/348fd45065620f20080299774f37f45233ef8f6b -> trunk/348fd45065620f20080299774f37f45233ef8f6b 2025-07-17T08:34:52.0176485Z * [new tag] trunk/3490a4f906aeaa867a390c2e97b357f17b01fcf2 -> trunk/3490a4f906aeaa867a390c2e97b357f17b01fcf2 2025-07-17T08:34:52.0178608Z * [new tag] trunk/34c8033fd3dea404db1706e264761912d095f723 -> trunk/34c8033fd3dea404db1706e264761912d095f723 2025-07-17T08:34:52.0180957Z * [new tag] trunk/34d8e64ef64d88324092a2028884c54c13e086b3 -> trunk/34d8e64ef64d88324092a2028884c54c13e086b3 2025-07-17T08:34:52.0183213Z * [new tag] trunk/35321b2ad6e86a58ca5d5be90ad8edce5945e16c -> trunk/35321b2ad6e86a58ca5d5be90ad8edce5945e16c 2025-07-17T08:34:52.0185442Z * [new tag] trunk/3580b8dde44d8bf4f229537ae9897ddd5e70f5db -> trunk/3580b8dde44d8bf4f229537ae9897ddd5e70f5db 2025-07-17T08:34:52.0187760Z * [new tag] trunk/3584e84c2434a2681b7288bcbd3bdb163c793df5 -> trunk/3584e84c2434a2681b7288bcbd3bdb163c793df5 2025-07-17T08:34:52.0190158Z * [new tag] trunk/3596c0c77f28c74e6cfb50769402ed52dbe10c67 -> trunk/3596c0c77f28c74e6cfb50769402ed52dbe10c67 2025-07-17T08:34:52.0192427Z * [new tag] trunk/359e8f5d691cdbbc54beafea25c06fc0ce321ad6 -> trunk/359e8f5d691cdbbc54beafea25c06fc0ce321ad6 2025-07-17T08:34:52.0196450Z * [new tag] trunk/35d03398e511fa0921c9db928c661dc9531ff2fc -> trunk/35d03398e511fa0921c9db928c661dc9531ff2fc 2025-07-17T08:34:52.0198695Z * [new tag] trunk/35e44067c4d9cc9be2652c0b9098885c5a321029 -> trunk/35e44067c4d9cc9be2652c0b9098885c5a321029 2025-07-17T08:34:52.0201077Z * [new tag] trunk/35e8f2593cdb72036c88a4b8d01487747204ba03 -> trunk/35e8f2593cdb72036c88a4b8d01487747204ba03 2025-07-17T08:34:52.0203263Z * [new tag] trunk/35ecd7c2d44a4e370e21ddab76b6c28266188846 -> trunk/35ecd7c2d44a4e370e21ddab76b6c28266188846 2025-07-17T08:34:52.0205476Z * [new tag] trunk/3608737347bcbb1ec1e23d85a29811abfd75c7a1 -> trunk/3608737347bcbb1ec1e23d85a29811abfd75c7a1 2025-07-17T08:34:52.0207717Z * [new tag] trunk/3644b41a7ccda62c066d525d956a08c9eac686c0 -> trunk/3644b41a7ccda62c066d525d956a08c9eac686c0 2025-07-17T08:34:52.0210046Z * [new tag] trunk/365ce465f393a6426b4ab3148da9a92445bf61d3 -> trunk/365ce465f393a6426b4ab3148da9a92445bf61d3 2025-07-17T08:34:52.0212289Z * [new tag] trunk/3684be056d9af667400ba071a116be8b1112bba8 -> trunk/3684be056d9af667400ba071a116be8b1112bba8 2025-07-17T08:34:52.0214583Z * [new tag] trunk/36a722e20d081c1a5a6df417d0f8333f6c082476 -> trunk/36a722e20d081c1a5a6df417d0f8333f6c082476 2025-07-17T08:34:52.0217018Z * [new tag] trunk/36bf81e36396c5f08eb8cab7f7858555282ca6c2 -> trunk/36bf81e36396c5f08eb8cab7f7858555282ca6c2 2025-07-17T08:34:52.0219283Z * [new tag] trunk/36dd598bdac5c665e46f05d00a38d6863a99615f -> trunk/36dd598bdac5c665e46f05d00a38d6863a99615f 2025-07-17T08:34:52.0221385Z * [new tag] trunk/36f7a027b595083f3a3761b407120b3f1f4e8634 -> trunk/36f7a027b595083f3a3761b407120b3f1f4e8634 2025-07-17T08:34:52.0223617Z * [new tag] trunk/36fd1ac9324429c095f8fbc5f6d2bd4b71f18d61 -> trunk/36fd1ac9324429c095f8fbc5f6d2bd4b71f18d61 2025-07-17T08:34:52.0225910Z * [new tag] trunk/370fc49dde9e8635957d9a910d43154085264225 -> trunk/370fc49dde9e8635957d9a910d43154085264225 2025-07-17T08:34:52.0228257Z * [new tag] trunk/376c16703c7f55a7eab9d5751bfc0e91af26a735 -> trunk/376c16703c7f55a7eab9d5751bfc0e91af26a735 2025-07-17T08:34:52.0230518Z * [new tag] trunk/378c121d5e1c8d2c24124f0c0b6cf08a1341e9d9 -> trunk/378c121d5e1c8d2c24124f0c0b6cf08a1341e9d9 2025-07-17T08:34:52.0232861Z * [new tag] trunk/37ccc532f75d8537c5b5d74b1c92a7f20df96353 -> trunk/37ccc532f75d8537c5b5d74b1c92a7f20df96353 2025-07-17T08:34:52.0235814Z * [new tag] trunk/380e30a723c1fb3530cad12ac44706412df7aa71 -> trunk/380e30a723c1fb3530cad12ac44706412df7aa71 2025-07-17T08:34:52.0238245Z * [new tag] trunk/3819584f12e2a46463a9799612ace1f59d76b9cd -> trunk/3819584f12e2a46463a9799612ace1f59d76b9cd 2025-07-17T08:34:52.0240182Z * [new tag] trunk/382598ef872b2afb9a03f8d88277a6c2edeb507f -> trunk/382598ef872b2afb9a03f8d88277a6c2edeb507f 2025-07-17T08:34:52.0242443Z * [new tag] trunk/382c6190c1329e96e71eef21a19737a3eda0040b -> trunk/382c6190c1329e96e71eef21a19737a3eda0040b 2025-07-17T08:34:52.0244608Z * [new tag] trunk/38371f693b07a485705119407da2e5dc64cec4eb -> trunk/38371f693b07a485705119407da2e5dc64cec4eb 2025-07-17T08:34:52.0247020Z * [new tag] trunk/38410cf9b57079f3360c1e79601973a01cb2588c -> trunk/38410cf9b57079f3360c1e79601973a01cb2588c 2025-07-17T08:34:52.0249132Z * [new tag] trunk/3863bbb55b38985c7d64c8a0be7beb2005a9cc07 -> trunk/3863bbb55b38985c7d64c8a0be7beb2005a9cc07 2025-07-17T08:34:52.0251374Z * [new tag] trunk/386aa7200324519ea9a8eff5eb1b3c0517756d24 -> trunk/386aa7200324519ea9a8eff5eb1b3c0517756d24 2025-07-17T08:34:52.0253692Z * [new tag] trunk/386bc9e2e990a6ac29bb90ba97b71c5b85e11080 -> trunk/386bc9e2e990a6ac29bb90ba97b71c5b85e11080 2025-07-17T08:34:52.0255965Z * [new tag] trunk/38757d94f1b3d65295b5ca2d7527ea0d582a5d3b -> trunk/38757d94f1b3d65295b5ca2d7527ea0d582a5d3b 2025-07-17T08:34:52.0258221Z * [new tag] trunk/38bfd462b8fb035de08a8c0d2b7b13eba78ee870 -> trunk/38bfd462b8fb035de08a8c0d2b7b13eba78ee870 2025-07-17T08:34:52.0260610Z * [new tag] trunk/38c4d05535f4e031a2ad5f39a4a415cc3a919cfb -> trunk/38c4d05535f4e031a2ad5f39a4a415cc3a919cfb 2025-07-17T08:34:52.0263175Z * [new tag] trunk/38e1e5d54ce42d42f8920ff46d9102800b94482d -> trunk/38e1e5d54ce42d42f8920ff46d9102800b94482d 2025-07-17T08:34:52.0265341Z * [new tag] trunk/38e5e81e55fc5d85d6cf8a83c96c88578995e3fe -> trunk/38e5e81e55fc5d85d6cf8a83c96c88578995e3fe 2025-07-17T08:34:52.0267582Z * [new tag] trunk/391473cca0b0fa9667ad54c972e7b63bc589cd6d -> trunk/391473cca0b0fa9667ad54c972e7b63bc589cd6d 2025-07-17T08:34:52.0270171Z * [new tag] trunk/39270430c9e4dc7010a9754f522774d17ae9d578 -> trunk/39270430c9e4dc7010a9754f522774d17ae9d578 2025-07-17T08:34:52.0272402Z * [new tag] trunk/39456edbbad97a09e6507792e1df3ee1f2a16f98 -> trunk/39456edbbad97a09e6507792e1df3ee1f2a16f98 2025-07-17T08:34:52.0274704Z * [new tag] trunk/398fca9dcfcb9f90eae9258d728fab71f5b59db2 -> trunk/398fca9dcfcb9f90eae9258d728fab71f5b59db2 2025-07-17T08:34:52.0277054Z * [new tag] trunk/39a8f66d5939e892bcb07ef97462af47d3201491 -> trunk/39a8f66d5939e892bcb07ef97462af47d3201491 2025-07-17T08:34:52.0279107Z * [new tag] trunk/39ac189808c61588f3594dbc2fc1d69bb6194c47 -> trunk/39ac189808c61588f3594dbc2fc1d69bb6194c47 2025-07-17T08:34:52.0281251Z * [new tag] trunk/39b71d11fc2dd9b4da6d23a34eb29aefbb1df672 -> trunk/39b71d11fc2dd9b4da6d23a34eb29aefbb1df672 2025-07-17T08:34:52.0283546Z * [new tag] trunk/39c605e8b3c13e7f6b5b1bbbe3f4060e24b2a3e4 -> trunk/39c605e8b3c13e7f6b5b1bbbe3f4060e24b2a3e4 2025-07-17T08:34:52.0285796Z * [new tag] trunk/3a43dba21ff5d01bb8b259af3839c90e447c6ec0 -> trunk/3a43dba21ff5d01bb8b259af3839c90e447c6ec0 2025-07-17T08:34:52.0288029Z * [new tag] trunk/3a5677a380c79810cc370dce6ef79b0871ddcf8c -> trunk/3a5677a380c79810cc370dce6ef79b0871ddcf8c 2025-07-17T08:34:52.0290259Z * [new tag] trunk/3a7ff829c597ef26a271d52ae2a3c8354b2f447e -> trunk/3a7ff829c597ef26a271d52ae2a3c8354b2f447e 2025-07-17T08:34:52.0292500Z * [new tag] trunk/3b4b5f8d474a9c664f622012c6a61414ea7799a9 -> trunk/3b4b5f8d474a9c664f622012c6a61414ea7799a9 2025-07-17T08:34:52.0294930Z * [new tag] trunk/3b6569b1ef4b9ff25f5b75fe0a216d6d084d573f -> trunk/3b6569b1ef4b9ff25f5b75fe0a216d6d084d573f 2025-07-17T08:34:52.0297157Z * [new tag] trunk/3b7c5e6fa5c0be64ddde3a8edc5bcdc10390f1e3 -> trunk/3b7c5e6fa5c0be64ddde3a8edc5bcdc10390f1e3 2025-07-17T08:34:52.0299256Z * [new tag] trunk/3bc6bdc8660c052d932f550d5734da6f801c2630 -> trunk/3bc6bdc8660c052d932f550d5734da6f801c2630 2025-07-17T08:34:52.0301618Z * [new tag] trunk/3bdd5ae334b85a114a2b62fa17dcf204413eda32 -> trunk/3bdd5ae334b85a114a2b62fa17dcf204413eda32 2025-07-17T08:34:52.0303849Z * [new tag] trunk/3beb915004f4e26b1e7c5e7692e6e8ca9b75de46 -> trunk/3beb915004f4e26b1e7c5e7692e6e8ca9b75de46 2025-07-17T08:34:52.0306126Z * [new tag] trunk/3bec588bf5c4eda9a4d42ae4c25e6f87af7f078c -> trunk/3bec588bf5c4eda9a4d42ae4c25e6f87af7f078c 2025-07-17T08:34:52.0308403Z * [new tag] trunk/3c2324c64ac6b7497d630788a66c82705bbb044e -> trunk/3c2324c64ac6b7497d630788a66c82705bbb044e 2025-07-17T08:34:52.0310672Z * [new tag] trunk/3c7079959c8e61d7acb4f704a0ecf74c61425c2e -> trunk/3c7079959c8e61d7acb4f704a0ecf74c61425c2e 2025-07-17T08:34:52.0312944Z * [new tag] trunk/3c72b9fd8feed4588a040bc681ffe83cc7acd26d -> trunk/3c72b9fd8feed4588a040bc681ffe83cc7acd26d 2025-07-17T08:34:52.0315341Z * [new tag] trunk/3cb11877aa30c04be7ffa9b4ca1722f1270a5828 -> trunk/3cb11877aa30c04be7ffa9b4ca1722f1270a5828 2025-07-17T08:34:52.0317475Z * [new tag] trunk/3cbae6dde8f2a9c3b4a3e4e079e97307c3aa52d8 -> trunk/3cbae6dde8f2a9c3b4a3e4e079e97307c3aa52d8 2025-07-17T08:34:52.0319792Z * [new tag] trunk/3d06ff82a84a118f0ed246864d4fc01ac4726328 -> trunk/3d06ff82a84a118f0ed246864d4fc01ac4726328 2025-07-17T08:34:52.0322222Z * [new tag] trunk/3d595fd5595f38bb5ed3d390dc50e1715e173ad6 -> trunk/3d595fd5595f38bb5ed3d390dc50e1715e173ad6 2025-07-17T08:34:52.0324473Z * [new tag] trunk/3d82a1dfb59fa5e248f7499a7ecdf784d4f61c0e -> trunk/3d82a1dfb59fa5e248f7499a7ecdf784d4f61c0e 2025-07-17T08:34:52.0326658Z * [new tag] trunk/3dabc351bb5581f69825eee6b24fbac9f9260241 -> trunk/3dabc351bb5581f69825eee6b24fbac9f9260241 2025-07-17T08:34:52.0328895Z * [new tag] trunk/3dd872e6d53560933d8d7fc11357617746d37168 -> trunk/3dd872e6d53560933d8d7fc11357617746d37168 2025-07-17T08:34:52.0331156Z * [new tag] trunk/3dda80e990121eaf156014fffe6e2a4602c8b195 -> trunk/3dda80e990121eaf156014fffe6e2a4602c8b195 2025-07-17T08:34:52.0333405Z * [new tag] trunk/3df6360e8c956edd25453b21f6b7f56e0366dcb4 -> trunk/3df6360e8c956edd25453b21f6b7f56e0366dcb4 2025-07-17T08:34:52.0335732Z * [new tag] trunk/3e131f7779af03b462f0598afb2569092b57c840 -> trunk/3e131f7779af03b462f0598afb2569092b57c840 2025-07-17T08:34:52.0338124Z * [new tag] trunk/3e38feb05fffdf5b181a1f4c7a6f43b00ef1c526 -> trunk/3e38feb05fffdf5b181a1f4c7a6f43b00ef1c526 2025-07-17T08:34:52.0340382Z * [new tag] trunk/3e56a9cdfb98a9b06568ee54e3157c800d98a17e -> trunk/3e56a9cdfb98a9b06568ee54e3157c800d98a17e 2025-07-17T08:34:52.0342649Z * [new tag] trunk/3eb7084f7a3657f4ba6626aca63721a0020f1bd7 -> trunk/3eb7084f7a3657f4ba6626aca63721a0020f1bd7 2025-07-17T08:34:52.0344881Z * [new tag] trunk/3ed4384f5b4bb7ae7d12298632a258385a51446e -> trunk/3ed4384f5b4bb7ae7d12298632a258385a51446e 2025-07-17T08:34:52.0347178Z * [new tag] trunk/3ee75b7eacef6758db602e87287ef9574609b327 -> trunk/3ee75b7eacef6758db602e87287ef9574609b327 2025-07-17T08:34:52.0349414Z * [new tag] trunk/3ee8828c87ce6186607c3d3ab3852518fca49228 -> trunk/3ee8828c87ce6186607c3d3ab3852518fca49228 2025-07-17T08:34:52.0351701Z * [new tag] trunk/3efb22e09111b92bedb01b2a8385c789fe69090a -> trunk/3efb22e09111b92bedb01b2a8385c789fe69090a 2025-07-17T08:34:52.0354036Z * [new tag] trunk/3effe0c293219b00a0eae7e139fe2d9aed84bc03 -> trunk/3effe0c293219b00a0eae7e139fe2d9aed84bc03 2025-07-17T08:34:52.0356549Z * [new tag] trunk/3f29642ecf039129032cc61c986d7b62807163c0 -> trunk/3f29642ecf039129032cc61c986d7b62807163c0 2025-07-17T08:34:52.0358824Z * [new tag] trunk/3f569f9af77d51d0328f16434e64252756681daa -> trunk/3f569f9af77d51d0328f16434e64252756681daa 2025-07-17T08:34:52.0361062Z * [new tag] trunk/3f65e38b73cb8f0d0b4fb2e9a6671085d371ec40 -> trunk/3f65e38b73cb8f0d0b4fb2e9a6671085d371ec40 2025-07-17T08:34:52.0363411Z * [new tag] trunk/3f69e3b3a07efd5e9aa86d37773bf2a24dc0ea70 -> trunk/3f69e3b3a07efd5e9aa86d37773bf2a24dc0ea70 2025-07-17T08:34:52.0365555Z * [new tag] trunk/3f83e3eeca0645f4b2cd16fa7d5a591e9cf810d4 -> trunk/3f83e3eeca0645f4b2cd16fa7d5a591e9cf810d4 2025-07-17T08:34:52.0367957Z * [new tag] trunk/3f920f3d8f5bd15d2222758f21f9a5d36e4dad1f -> trunk/3f920f3d8f5bd15d2222758f21f9a5d36e4dad1f 2025-07-17T08:34:52.0370221Z * [new tag] trunk/3fd84a8592a2a87d04f0e3f236a92605e2be12df -> trunk/3fd84a8592a2a87d04f0e3f236a92605e2be12df 2025-07-17T08:34:52.0372489Z * [new tag] trunk/400f439670ff8859c17a870eb83dc5cae5a9c2e4 -> trunk/400f439670ff8859c17a870eb83dc5cae5a9c2e4 2025-07-17T08:34:52.0374842Z * [new tag] trunk/402ae09e41005ebff686e97e2f120a6b79e2afb7 -> trunk/402ae09e41005ebff686e97e2f120a6b79e2afb7 2025-07-17T08:34:52.0377146Z * [new tag] trunk/404008e3efdabeaf5b140a3aff77131461c33a0a -> trunk/404008e3efdabeaf5b140a3aff77131461c33a0a 2025-07-17T08:34:52.0379416Z * [new tag] trunk/4048a144abf82041f9b6653d9d4aa600a6a5409a -> trunk/4048a144abf82041f9b6653d9d4aa600a6a5409a 2025-07-17T08:34:52.0381974Z * [new tag] trunk/408d9884b07cf7268961bae7138a6436916d4a43 -> trunk/408d9884b07cf7268961bae7138a6436916d4a43 2025-07-17T08:34:52.0384438Z * [new tag] trunk/40a785103cf94a1dbc3e0e43d1ed6c41fb60bedb -> trunk/40a785103cf94a1dbc3e0e43d1ed6c41fb60bedb 2025-07-17T08:34:52.0386597Z * [new tag] trunk/40d02eb481670325ba80dcb42b40c3ebb347f1c7 -> trunk/40d02eb481670325ba80dcb42b40c3ebb347f1c7 2025-07-17T08:34:52.0388901Z * [new tag] trunk/40e39ae21f15f200828ab32826bed1a4b62532ba -> trunk/40e39ae21f15f200828ab32826bed1a4b62532ba 2025-07-17T08:34:52.0391167Z * [new tag] trunk/40fefe2871a5561b67ec268bab19db60e733f2d4 -> trunk/40fefe2871a5561b67ec268bab19db60e733f2d4 2025-07-17T08:34:52.0393387Z * [new tag] trunk/414ad470450c654d97e73bef704a7b596b5b4cbc -> trunk/414ad470450c654d97e73bef704a7b596b5b4cbc 2025-07-17T08:34:52.0395989Z * [new tag] trunk/415dfabe9b569b71098a2f874f3fc67ad2a4fc2e -> trunk/415dfabe9b569b71098a2f874f3fc67ad2a4fc2e 2025-07-17T08:34:52.0398217Z * [new tag] trunk/4162c0f70297818abd70f2fe7424dcd3cc9b2543 -> trunk/4162c0f70297818abd70f2fe7424dcd3cc9b2543 2025-07-17T08:34:52.0400424Z * [new tag] trunk/41910d7a94d2f3ffe06f7c0c30971c71b9c3b09e -> trunk/41910d7a94d2f3ffe06f7c0c30971c71b9c3b09e 2025-07-17T08:34:52.0402733Z * [new tag] trunk/41971335c98b0881e0784085096eceace575d563 -> trunk/41971335c98b0881e0784085096eceace575d563 2025-07-17T08:34:52.0405073Z * [new tag] trunk/41e8b826d07bbf707bc2c64a78b4dc15e8f3c358 -> trunk/41e8b826d07bbf707bc2c64a78b4dc15e8f3c358 2025-07-17T08:34:52.0407313Z * [new tag] trunk/41f6acef83d280a18909d58f8442e145c9e7ea6f -> trunk/41f6acef83d280a18909d58f8442e145c9e7ea6f 2025-07-17T08:34:52.0409608Z * [new tag] trunk/42015db6a9602f40250f9afc18a3fbad4ca4ef39 -> trunk/42015db6a9602f40250f9afc18a3fbad4ca4ef39 2025-07-17T08:34:52.0412057Z * [new tag] trunk/4237ee3c33b0edb01db96d8c1ecf8f5d2cf184bc -> trunk/4237ee3c33b0edb01db96d8c1ecf8f5d2cf184bc 2025-07-17T08:34:52.0414310Z * [new tag] trunk/4283d96bcdf9aaa4289985267186d74ba6534ee5 -> trunk/4283d96bcdf9aaa4289985267186d74ba6534ee5 2025-07-17T08:34:52.0416575Z * [new tag] trunk/42b48ee67229286127390000f103a11dfc8901f5 -> trunk/42b48ee67229286127390000f103a11dfc8901f5 2025-07-17T08:34:52.0418925Z * [new tag] trunk/42ff6a4a5c4e0d77bd18fcc5426622f1b8f20add -> trunk/42ff6a4a5c4e0d77bd18fcc5426622f1b8f20add 2025-07-17T08:34:52.0421186Z * [new tag] trunk/430cc1c636380a7d50652df646274f7008a76747 -> trunk/430cc1c636380a7d50652df646274f7008a76747 2025-07-17T08:34:52.0423453Z * [new tag] trunk/4311aea5e7568c9956a9f0d694f1980c387b6a37 -> trunk/4311aea5e7568c9956a9f0d694f1980c387b6a37 2025-07-17T08:34:52.0425708Z * [new tag] trunk/433a2471023f77a4e62e880ef59f22caf939c227 -> trunk/433a2471023f77a4e62e880ef59f22caf939c227 2025-07-17T08:34:52.0428003Z * [new tag] trunk/43523bf1682c2926a84a1f65f00fabd3d34db4f2 -> trunk/43523bf1682c2926a84a1f65f00fabd3d34db4f2 2025-07-17T08:34:52.0430185Z * [new tag] trunk/43a09189c68fe02bd9d8433c4a144ffc9bbf895c -> trunk/43a09189c68fe02bd9d8433c4a144ffc9bbf895c 2025-07-17T08:34:52.0432537Z * [new tag] trunk/43f72163275ce1503b1de1480642c61d20e6158d -> trunk/43f72163275ce1503b1de1480642c61d20e6158d 2025-07-17T08:34:52.0436529Z * [new tag] trunk/442aca44d603ae6c2b7d2aa2190cc91f970c4202 -> trunk/442aca44d603ae6c2b7d2aa2190cc91f970c4202 2025-07-17T08:34:52.0438892Z * [new tag] trunk/44303caabfa0bdc5688d693d491ba7c3fdb40f3a -> trunk/44303caabfa0bdc5688d693d491ba7c3fdb40f3a 2025-07-17T08:34:52.0441212Z * [new tag] trunk/443b5b43c3d40b89bbb5db70f3f05a44b2a4ec66 -> trunk/443b5b43c3d40b89bbb5db70f3f05a44b2a4ec66 2025-07-17T08:34:52.0443582Z * [new tag] trunk/4486a6dbfd65ef490cfe73e0630929e85f61ee16 -> trunk/4486a6dbfd65ef490cfe73e0630929e85f61ee16 2025-07-17T08:34:52.0445843Z * [new tag] trunk/4491326fb0c0e67eca1598ae33c41cdfced2cd33 -> trunk/4491326fb0c0e67eca1598ae33c41cdfced2cd33 2025-07-17T08:34:52.0447940Z * [new tag] trunk/44a5f93462bd4aa4f36ae2573685ed7ba458c4c9 -> trunk/44a5f93462bd4aa4f36ae2573685ed7ba458c4c9 2025-07-17T08:34:52.0450269Z * [new tag] trunk/44d0800d60e78fef8ab332e307c3134e3c276ba4 -> trunk/44d0800d60e78fef8ab332e307c3134e3c276ba4 2025-07-17T08:34:52.0452541Z * [new tag] trunk/44df7cf28dd37a4d900eb6b2f78dabf72b209d9b -> trunk/44df7cf28dd37a4d900eb6b2f78dabf72b209d9b 2025-07-17T08:34:52.0454770Z * [new tag] trunk/44f5b9312290866584088e2e42228c484f669faf -> trunk/44f5b9312290866584088e2e42228c484f669faf 2025-07-17T08:34:52.0456898Z * [new tag] trunk/4500a4aa50141ed30e093ef8491b30d1d1287348 -> trunk/4500a4aa50141ed30e093ef8491b30d1d1287348 2025-07-17T08:34:52.0459139Z * [new tag] trunk/451b525bf0cb8840adbde1b6280c7b199f733ced -> trunk/451b525bf0cb8840adbde1b6280c7b199f733ced 2025-07-17T08:34:52.0461424Z * [new tag] trunk/45382b284d03015f3e6eb83a959f0e54a3ebd688 -> trunk/45382b284d03015f3e6eb83a959f0e54a3ebd688 2025-07-17T08:34:52.0463718Z * [new tag] trunk/453bc9fbdfb97fef925477cbbf4948e93fd22756 -> trunk/453bc9fbdfb97fef925477cbbf4948e93fd22756 2025-07-17T08:34:52.0465965Z * [new tag] trunk/45596ec58f5bc5489911ee932c2a55a7ff40d0a0 -> trunk/45596ec58f5bc5489911ee932c2a55a7ff40d0a0 2025-07-17T08:34:52.0468258Z * [new tag] trunk/455dfd258980294f0745bd90aee12a323e37224d -> trunk/455dfd258980294f0745bd90aee12a323e37224d 2025-07-17T08:34:52.0470566Z * [new tag] trunk/456b7451c78096fd512c60e26cc618386a251bc7 -> trunk/456b7451c78096fd512c60e26cc618386a251bc7 2025-07-17T08:34:52.0472830Z * [new tag] trunk/456f40cb09c6e5e0cd268d1601b89fc5ac1987ea -> trunk/456f40cb09c6e5e0cd268d1601b89fc5ac1987ea 2025-07-17T08:34:52.0475294Z * [new tag] trunk/4574b39aa45f0250cba04aa3cb053a686c4888b2 -> trunk/4574b39aa45f0250cba04aa3cb053a686c4888b2 2025-07-17T08:34:52.0477632Z * [new tag] trunk/457dd79927db6ae7f1c53f36768b7ac93b61e507 -> trunk/457dd79927db6ae7f1c53f36768b7ac93b61e507 2025-07-17T08:34:52.0479864Z * [new tag] trunk/4585c33e74079af8e9067bf39970b93c3f13629f -> trunk/4585c33e74079af8e9067bf39970b93c3f13629f 2025-07-17T08:34:52.0482087Z * [new tag] trunk/458cc7213baf01ab742a2250c7a31f7796fcb1ef -> trunk/458cc7213baf01ab742a2250c7a31f7796fcb1ef 2025-07-17T08:34:52.0484364Z * [new tag] trunk/45c5a232373cac1b1524259f20758bfb56dde5f2 -> trunk/45c5a232373cac1b1524259f20758bfb56dde5f2 2025-07-17T08:34:52.0486721Z * [new tag] trunk/4609699bfd440d1fe603c5ad9a942e45410094c1 -> trunk/4609699bfd440d1fe603c5ad9a942e45410094c1 2025-07-17T08:34:52.0489080Z * [new tag] trunk/4628f1b7a9313df98346d582d9a789661a823e27 -> trunk/4628f1b7a9313df98346d582d9a789661a823e27 2025-07-17T08:34:52.0491413Z * [new tag] trunk/463fe36532bc165b46c12cef4948f039434e0924 -> trunk/463fe36532bc165b46c12cef4948f039434e0924 2025-07-17T08:34:52.0493630Z * [new tag] trunk/4657a84bc55b6ce12f21706de2b90e1d43784f57 -> trunk/4657a84bc55b6ce12f21706de2b90e1d43784f57 2025-07-17T08:34:52.0495901Z * [new tag] trunk/46915b13614dbac90724d0f1802b8e0db037c9e4 -> trunk/46915b13614dbac90724d0f1802b8e0db037c9e4 2025-07-17T08:34:52.0498196Z * [new tag] trunk/473208cb18d543e8f968918a6b3c9defa8a4ae10 -> trunk/473208cb18d543e8f968918a6b3c9defa8a4ae10 2025-07-17T08:34:52.0500444Z * [new tag] trunk/476874b37fff42a46d25dfac720ef4c71ec74fe0 -> trunk/476874b37fff42a46d25dfac720ef4c71ec74fe0 2025-07-17T08:34:52.0502918Z * [new tag] trunk/4781b0ee6057fdf2a82d2f9ed30299f219267a71 -> trunk/4781b0ee6057fdf2a82d2f9ed30299f219267a71 2025-07-17T08:34:52.0505270Z * [new tag] trunk/4781d72faa6b72bf96fde9bedfca06a5285eebfb -> trunk/4781d72faa6b72bf96fde9bedfca06a5285eebfb 2025-07-17T08:34:52.0507681Z * [new tag] trunk/47c8810b5275179833d6b33ca3d70922f485272c -> trunk/47c8810b5275179833d6b33ca3d70922f485272c 2025-07-17T08:34:52.0510706Z * [new tag] trunk/47f10d0ad0dda281c886ff08ac2f938207027316 -> trunk/47f10d0ad0dda281c886ff08ac2f938207027316 2025-07-17T08:34:52.0513068Z * [new tag] trunk/4805a6ead6f1e7f32351056e2602be4e908f69b7 -> trunk/4805a6ead6f1e7f32351056e2602be4e908f69b7 2025-07-17T08:34:52.0515492Z * [new tag] trunk/48315181c75e43cab5957197d42e053d66b3fe1c -> trunk/48315181c75e43cab5957197d42e053d66b3fe1c 2025-07-17T08:34:52.0517883Z * [new tag] trunk/4851863e3f98e1ffba1d5801f3533031ef4f905d -> trunk/4851863e3f98e1ffba1d5801f3533031ef4f905d 2025-07-17T08:34:52.0520107Z * [new tag] trunk/48560eef80e97e855cbb8e2814acefe8f5cc6fbd -> trunk/48560eef80e97e855cbb8e2814acefe8f5cc6fbd 2025-07-17T08:34:52.0522360Z * [new tag] trunk/4886ba64dceb24b8a0444e574abb907a18fdeee9 -> trunk/4886ba64dceb24b8a0444e574abb907a18fdeee9 2025-07-17T08:34:52.0524687Z * [new tag] trunk/48921721d8ed32a2c7f80605b434f9fcb2f948b4 -> trunk/48921721d8ed32a2c7f80605b434f9fcb2f948b4 2025-07-17T08:34:52.0526990Z * [new tag] trunk/48de3da2539cecaee14af8e3841c133c9c0c0f1c -> trunk/48de3da2539cecaee14af8e3841c133c9c0c0f1c 2025-07-17T08:34:52.0529073Z * [new tag] trunk/48e7b62d3abda4a3eac2f538c9f35e56093c4ece -> trunk/48e7b62d3abda4a3eac2f538c9f35e56093c4ece 2025-07-17T08:34:52.0531354Z * [new tag] trunk/4918502d2e685270cea9f1fd733a3414ac1ca5d7 -> trunk/4918502d2e685270cea9f1fd733a3414ac1ca5d7 2025-07-17T08:34:52.0533593Z * [new tag] trunk/493bd625e252dea02e871346beaa49745b4b2663 -> trunk/493bd625e252dea02e871346beaa49745b4b2663 2025-07-17T08:34:52.0535957Z * [new tag] trunk/493f42a5417a8ac33d1070478d1869ea4f207b5b -> trunk/493f42a5417a8ac33d1070478d1869ea4f207b5b 2025-07-17T08:34:52.0538200Z * [new tag] trunk/495c317005ad656dee752b6ad7bd5541b04e388e -> trunk/495c317005ad656dee752b6ad7bd5541b04e388e 2025-07-17T08:34:52.0540410Z * [new tag] trunk/49888e6be0dcc00ba546746425893bb17c287248 -> trunk/49888e6be0dcc00ba546746425893bb17c287248 2025-07-17T08:34:52.0543881Z * [new tag] trunk/49ee1e7106db00778345efd7c70cc12d2ca6a91a -> trunk/49ee1e7106db00778345efd7c70cc12d2ca6a91a 2025-07-17T08:34:52.0546002Z * [new tag] trunk/4a26bb8a12ba5cb437ed5b7f035b5b533e07549e -> trunk/4a26bb8a12ba5cb437ed5b7f035b5b533e07549e 2025-07-17T08:34:52.0548322Z * [new tag] trunk/4a4cac0cefea3661cc69cfdafdba64832ee0841a -> trunk/4a4cac0cefea3661cc69cfdafdba64832ee0841a 2025-07-17T08:34:52.0550591Z * [new tag] trunk/4a80ddfbe70bf6b75acc3177e5d2095b285da841 -> trunk/4a80ddfbe70bf6b75acc3177e5d2095b285da841 2025-07-17T08:34:52.0552893Z * [new tag] trunk/4a8f5e752beb5a6809ba866c83f32dd464a47bfd -> trunk/4a8f5e752beb5a6809ba866c83f32dd464a47bfd 2025-07-17T08:34:52.0555369Z * [new tag] trunk/4a954fc1857ff41be604e8cae68908d41124b419 -> trunk/4a954fc1857ff41be604e8cae68908d41124b419 2025-07-17T08:34:52.0557754Z * [new tag] trunk/4ab4d29cbee1f90daf14444d5ca3a53653c7784d -> trunk/4ab4d29cbee1f90daf14444d5ca3a53653c7784d 2025-07-17T08:34:52.0559975Z * [new tag] trunk/4b11428cb5b3d97f3068a2dc4c55cee6ddd41979 -> trunk/4b11428cb5b3d97f3068a2dc4c55cee6ddd41979 2025-07-17T08:34:52.0562335Z * [new tag] trunk/4b4c2a7b1dfd88313801878c5b4e3855fe5232df -> trunk/4b4c2a7b1dfd88313801878c5b4e3855fe5232df 2025-07-17T08:34:52.0565231Z * [new tag] trunk/4b55871e06d6bad54eac45e45a9af615d758a39f -> trunk/4b55871e06d6bad54eac45e45a9af615d758a39f 2025-07-17T08:34:52.0567378Z * [new tag] trunk/4b6cbf528b8f7b95e02e3a0233945e0bb8b00f44 -> trunk/4b6cbf528b8f7b95e02e3a0233945e0bb8b00f44 2025-07-17T08:34:52.0569956Z * [new tag] trunk/4b9a6f7211123511e856ac8c8524bc332a741241 -> trunk/4b9a6f7211123511e856ac8c8524bc332a741241 2025-07-17T08:34:52.0572097Z * [new tag] trunk/4bb936d8b77c959efbff0bdb68f63e5a5faf60a9 -> trunk/4bb936d8b77c959efbff0bdb68f63e5a5faf60a9 2025-07-17T08:34:52.0574402Z * [new tag] trunk/4bc3e4b497948af0a48d08b427339efc8960fedb -> trunk/4bc3e4b497948af0a48d08b427339efc8960fedb 2025-07-17T08:34:52.0576674Z * [new tag] trunk/4bd18e31e5a38d0e84ce915b1fa124058c6373fa -> trunk/4bd18e31e5a38d0e84ce915b1fa124058c6373fa 2025-07-17T08:34:52.0578918Z * [new tag] trunk/4c0091fda65b714fa73671a15e379f814af153e0 -> trunk/4c0091fda65b714fa73671a15e379f814af153e0 2025-07-17T08:34:52.0581483Z * [new tag] trunk/4c0aa37dda605e7cf3372c71940d7e3a93ecef8d -> trunk/4c0aa37dda605e7cf3372c71940d7e3a93ecef8d 2025-07-17T08:34:52.0583722Z * [new tag] trunk/4c3da611c2a80fa5073b164974d7edc0577b9ca7 -> trunk/4c3da611c2a80fa5073b164974d7edc0577b9ca7 2025-07-17T08:34:52.0586047Z * [new tag] trunk/4c59edf0c5838087cbe09c67c6e0e776467a5f2d -> trunk/4c59edf0c5838087cbe09c67c6e0e776467a5f2d 2025-07-17T08:34:52.0588284Z * [new tag] trunk/4c8eb65efb147cd263fc02f5588683f530363a0f -> trunk/4c8eb65efb147cd263fc02f5588683f530363a0f 2025-07-17T08:34:52.0590559Z * [new tag] trunk/4cbbc8b4583e9dac4af7e8d1d8535546de21178c -> trunk/4cbbc8b4583e9dac4af7e8d1d8535546de21178c 2025-07-17T08:34:52.0592808Z * [new tag] trunk/4cc13c4af6dbd5aa36545ef577d0bee517cf8c57 -> trunk/4cc13c4af6dbd5aa36545ef577d0bee517cf8c57 2025-07-17T08:34:52.0595219Z * [new tag] trunk/4cd6e96bf0d13f168772f6ad44267b1ac7632a97 -> trunk/4cd6e96bf0d13f168772f6ad44267b1ac7632a97 2025-07-17T08:34:52.0597561Z * [new tag] trunk/4cdbdcdbcf2b7f7a52a5c61cb35064f7b4aa2e44 -> trunk/4cdbdcdbcf2b7f7a52a5c61cb35064f7b4aa2e44 2025-07-17T08:34:52.0599849Z * [new tag] trunk/4ce6e6ec8890a3f6ee604c9efb3ff153825ce575 -> trunk/4ce6e6ec8890a3f6ee604c9efb3ff153825ce575 2025-07-17T08:34:52.0602075Z * [new tag] trunk/4cfc0a320897b6ec75b85cfc07bc29009a824177 -> trunk/4cfc0a320897b6ec75b85cfc07bc29009a824177 2025-07-17T08:34:52.0604311Z * [new tag] trunk/4d055982e38f59fdb2a4c9d8855e58548bc42c12 -> trunk/4d055982e38f59fdb2a4c9d8855e58548bc42c12 2025-07-17T08:34:52.0606727Z * [new tag] trunk/4d3ecefda5a41df678fd68b020c521db95e9fbde -> trunk/4d3ecefda5a41df678fd68b020c521db95e9fbde 2025-07-17T08:34:52.0608957Z * [new tag] trunk/4d5d627e5ff3310318f9df80348ee419324228df -> trunk/4d5d627e5ff3310318f9df80348ee419324228df 2025-07-17T08:34:52.0611192Z * [new tag] trunk/4d9d884c3f5adc89b59d25d5a080498b76f9bb39 -> trunk/4d9d884c3f5adc89b59d25d5a080498b76f9bb39 2025-07-17T08:34:52.0613700Z * [new tag] trunk/4da98351b9e231bbbe83b5c590cf7d2ea382333a -> trunk/4da98351b9e231bbbe83b5c590cf7d2ea382333a 2025-07-17T08:34:52.0616099Z * [new tag] trunk/4dce5b71a0751aa4287c3a80adf07afc773ec4db -> trunk/4dce5b71a0751aa4287c3a80adf07afc773ec4db 2025-07-17T08:34:52.0618443Z * [new tag] trunk/4e13eca713c60ca63c1116823b99d2461a7422ef -> trunk/4e13eca713c60ca63c1116823b99d2461a7422ef 2025-07-17T08:34:52.0620687Z * [new tag] trunk/4e19477196547eb2e8157d6d132689373ffcf0fa -> trunk/4e19477196547eb2e8157d6d132689373ffcf0fa 2025-07-17T08:34:52.0622969Z * [new tag] trunk/4e8dd11be17a974e0ee5503b3061d4d47467844f -> trunk/4e8dd11be17a974e0ee5503b3061d4d47467844f 2025-07-17T08:34:52.0625253Z * [new tag] trunk/4ebd2690654ffe487c43941ebaea41b99914d8b7 -> trunk/4ebd2690654ffe487c43941ebaea41b99914d8b7 2025-07-17T08:34:52.0627536Z * [new tag] trunk/4ed1b03f7275075b1783d1deab946ced4ceba4d6 -> trunk/4ed1b03f7275075b1783d1deab946ced4ceba4d6 2025-07-17T08:34:52.0629976Z * [new tag] trunk/4ee4863232b9e07728d85254768bcba3aadc9b9a -> trunk/4ee4863232b9e07728d85254768bcba3aadc9b9a 2025-07-17T08:34:52.0632068Z * [new tag] trunk/4f36743f5eef2d9c40357eb5d8d8b1aeeacfbb2a -> trunk/4f36743f5eef2d9c40357eb5d8d8b1aeeacfbb2a 2025-07-17T08:34:52.0634389Z * [new tag] trunk/4f5b34427b57e8b876d12e6ce551f04a7c31cacf -> trunk/4f5b34427b57e8b876d12e6ce551f04a7c31cacf 2025-07-17T08:34:52.0637231Z * [new tag] trunk/4f5be5661240427fbbfbe1d137dce53ca32846b5 -> trunk/4f5be5661240427fbbfbe1d137dce53ca32846b5 2025-07-17T08:34:52.0639176Z * [new tag] trunk/4f70fbbd16d1f0d62af082246a95e56cffccc860 -> trunk/4f70fbbd16d1f0d62af082246a95e56cffccc860 2025-07-17T08:34:52.0641520Z * [new tag] trunk/4ff0e033c16aae064f05ee9300053cd304891673 -> trunk/4ff0e033c16aae064f05ee9300053cd304891673 2025-07-17T08:34:52.0643612Z * [new tag] trunk/4ff9b7fa3116b1c429e577830ac6e816734ad029 -> trunk/4ff9b7fa3116b1c429e577830ac6e816734ad029 2025-07-17T08:34:52.0645865Z * [new tag] trunk/502486d9466e81cda76c91d33eb869235480bee9 -> trunk/502486d9466e81cda76c91d33eb869235480bee9 2025-07-17T08:34:52.0647960Z * [new tag] trunk/503362d019b3782581492af7767945dbd75ca1c9 -> trunk/503362d019b3782581492af7767945dbd75ca1c9 2025-07-17T08:34:52.0650093Z * [new tag] trunk/508cdc4fc9f7a91b5fd3f20de38d673daebf1af3 -> trunk/508cdc4fc9f7a91b5fd3f20de38d673daebf1af3 2025-07-17T08:34:52.0652292Z * [new tag] trunk/50940270ae179134cd4f9072f04ffdd55daf808e -> trunk/50940270ae179134cd4f9072f04ffdd55daf808e 2025-07-17T08:34:52.0654642Z * [new tag] trunk/50b2069b61942e923528c94ccbbc8ab5e92c381e -> trunk/50b2069b61942e923528c94ccbbc8ab5e92c381e 2025-07-17T08:34:52.0656874Z * [new tag] trunk/50d8168c8b62990fda86398c9dee1dc8cbd6908d -> trunk/50d8168c8b62990fda86398c9dee1dc8cbd6908d 2025-07-17T08:34:52.0659249Z * [new tag] trunk/510c398a4f1dff9217938159d605290149358332 -> trunk/510c398a4f1dff9217938159d605290149358332 2025-07-17T08:34:52.0661646Z * [new tag] trunk/5116293f7eb587ada6076f3c3ea1711a0ec7ab4a -> trunk/5116293f7eb587ada6076f3c3ea1711a0ec7ab4a 2025-07-17T08:34:52.0663888Z * [new tag] trunk/5118a8f8a5a7906e26bcbd35370de3416b5cdab0 -> trunk/5118a8f8a5a7906e26bcbd35370de3416b5cdab0 2025-07-17T08:34:52.0666162Z * [new tag] trunk/51560797ce70aea353585d8381ee88c2e6c81075 -> trunk/51560797ce70aea353585d8381ee88c2e6c81075 2025-07-17T08:34:52.0668474Z * [new tag] trunk/517d2995e09603017f07f065c1ae5b8d25962cd2 -> trunk/517d2995e09603017f07f065c1ae5b8d25962cd2 2025-07-17T08:34:52.0670773Z * [new tag] trunk/51853b358e698f6a42c1eef045ab8fa766129ee1 -> trunk/51853b358e698f6a42c1eef045ab8fa766129ee1 2025-07-17T08:34:52.0673023Z * [new tag] trunk/51a708ffc679b13f99e4c7cf19bc00082a3266a6 -> trunk/51a708ffc679b13f99e4c7cf19bc00082a3266a6 2025-07-17T08:34:52.0676809Z * [new tag] trunk/51eb8e8f84bb9aa901cff17dd649e18b17a8908c -> trunk/51eb8e8f84bb9aa901cff17dd649e18b17a8908c 2025-07-17T08:34:52.0679352Z * [new tag] trunk/52214485747ceb0ea47ce5a09babb0b82b3282b8 -> trunk/52214485747ceb0ea47ce5a09babb0b82b3282b8 2025-07-17T08:34:52.0681590Z * [new tag] trunk/522a18bd6c094c766c0de9ef539682d1a3f04a15 -> trunk/522a18bd6c094c766c0de9ef539682d1a3f04a15 2025-07-17T08:34:52.0683795Z * [new tag] trunk/523b637cbeb69665072a2cf489ec1c5313b57670 -> trunk/523b637cbeb69665072a2cf489ec1c5313b57670 2025-07-17T08:34:52.0686205Z * [new tag] trunk/524e8270955788c53473497fd2cc16c5aa0e4c67 -> trunk/524e8270955788c53473497fd2cc16c5aa0e4c67 2025-07-17T08:34:52.0689312Z * [new tag] trunk/5264f8cd8d08272003298cdefe6bd60b1b8c80b4 -> trunk/5264f8cd8d08272003298cdefe6bd60b1b8c80b4 2025-07-17T08:34:52.0691587Z * [new tag] trunk/52772765e034622d1a86476e4bb19c28e3945f74 -> trunk/52772765e034622d1a86476e4bb19c28e3945f74 2025-07-17T08:34:52.0694057Z * [new tag] trunk/5285d1024376396a00fe750fe3d628c65dc26254 -> trunk/5285d1024376396a00fe750fe3d628c65dc26254 2025-07-17T08:34:52.0696168Z * [new tag] trunk/529e0357c6c4e74f8cd32c29198c5f1c9f6e329d -> trunk/529e0357c6c4e74f8cd32c29198c5f1c9f6e329d 2025-07-17T08:34:52.0698429Z * [new tag] trunk/52e4e41cbc36a5cf44395ff84ca2d069263560de -> trunk/52e4e41cbc36a5cf44395ff84ca2d069263560de 2025-07-17T08:34:52.0700740Z * [new tag] trunk/52edfb2cbcdded865645cc82bb1bb501fdcbdb52 -> trunk/52edfb2cbcdded865645cc82bb1bb501fdcbdb52 2025-07-17T08:34:52.0703005Z * [new tag] trunk/52f873adc23e7069ce95933d97ee514022cfbf7d -> trunk/52f873adc23e7069ce95933d97ee514022cfbf7d 2025-07-17T08:34:52.0705253Z * [new tag] trunk/53057fc16af2d381a61fe891b52ac8731ee9dfae -> trunk/53057fc16af2d381a61fe891b52ac8731ee9dfae 2025-07-17T08:34:52.0707490Z * [new tag] trunk/534c454e77ac0eefd52d63c60d42911e7f9617ea -> trunk/534c454e77ac0eefd52d63c60d42911e7f9617ea 2025-07-17T08:34:52.0709734Z * [new tag] trunk/53ab73090eb66d96e45ec134e41cf67266208954 -> trunk/53ab73090eb66d96e45ec134e41cf67266208954 2025-07-17T08:34:52.0712131Z * [new tag] trunk/53cd18f6b31bc2ea62985fda87e96aa17cd11bc1 -> trunk/53cd18f6b31bc2ea62985fda87e96aa17cd11bc1 2025-07-17T08:34:52.0714507Z * [new tag] trunk/53d06e18d9b165cb8aa0a5a3cbb6837ac3000c59 -> trunk/53d06e18d9b165cb8aa0a5a3cbb6837ac3000c59 2025-07-17T08:34:52.0717007Z * [new tag] trunk/53e0b9c3936176521ed8d71c00abd5b7499057c2 -> trunk/53e0b9c3936176521ed8d71c00abd5b7499057c2 2025-07-17T08:34:52.0719283Z * [new tag] trunk/541297daae63f74a90bac52c5db3540bf84bc971 -> trunk/541297daae63f74a90bac52c5db3540bf84bc971 2025-07-17T08:34:52.0721526Z * [new tag] trunk/5430990bd7d050f249b1cdf1f7d1016c4b17488d -> trunk/5430990bd7d050f249b1cdf1f7d1016c4b17488d 2025-07-17T08:34:52.0723942Z * [new tag] trunk/5435e7539930be7bd038683ce096038b30c5bb5f -> trunk/5435e7539930be7bd038683ce096038b30c5bb5f 2025-07-17T08:34:52.0726218Z * [new tag] trunk/545fbd58dc8b4f688f6b259cb671b3ec16a46810 -> trunk/545fbd58dc8b4f688f6b259cb671b3ec16a46810 2025-07-17T08:34:52.0728464Z * [new tag] trunk/54701a0c943245402fceeab0e55d7aa394303c20 -> trunk/54701a0c943245402fceeab0e55d7aa394303c20 2025-07-17T08:34:52.0730794Z * [new tag] trunk/5484890539823d9867c74209588abe095c9232a1 -> trunk/5484890539823d9867c74209588abe095c9232a1 2025-07-17T08:34:52.0733483Z * [new tag] trunk/548c9d8281d9b3d6d12e90c40b3387faf0e24c61 -> trunk/548c9d8281d9b3d6d12e90c40b3387faf0e24c61 2025-07-17T08:34:52.0735545Z * [new tag] trunk/54976bca103fcf2b5037cc0cd1b37c4639fcf779 -> trunk/54976bca103fcf2b5037cc0cd1b37c4639fcf779 2025-07-17T08:34:52.0737806Z * [new tag] trunk/54998c2daaf5b8919cf82367492dd3c5177ab935 -> trunk/54998c2daaf5b8919cf82367492dd3c5177ab935 2025-07-17T08:34:52.0740179Z * [new tag] trunk/54a4d34d100f4ebe45f486451967daba896e839c -> trunk/54a4d34d100f4ebe45f486451967daba896e839c 2025-07-17T08:34:52.0742400Z * [new tag] trunk/54a7e5b5983d237b324b50703bcb0919a6c4c296 -> trunk/54a7e5b5983d237b324b50703bcb0919a6c4c296 2025-07-17T08:34:52.0744714Z * [new tag] trunk/54b8087f638de57c1c93c4624d03f2f2c512b90e -> trunk/54b8087f638de57c1c93c4624d03f2f2c512b90e 2025-07-17T08:34:52.0747077Z * [new tag] trunk/55108074c0795be3b617d3b13b06794f63e1f8ca -> trunk/55108074c0795be3b617d3b13b06794f63e1f8ca 2025-07-17T08:34:52.0749433Z * [new tag] trunk/554b5680405e6197a985040ffe88157beb637450 -> trunk/554b5680405e6197a985040ffe88157beb637450 2025-07-17T08:34:52.0751816Z * [new tag] trunk/555f3562541992b66a550eca8e8740884b1247f8 -> trunk/555f3562541992b66a550eca8e8740884b1247f8 2025-07-17T08:34:52.0754252Z * [new tag] trunk/5596cefba623fcc0739f1e9568131a85973cf1f9 -> trunk/5596cefba623fcc0739f1e9568131a85973cf1f9 2025-07-17T08:34:52.0756528Z * [new tag] trunk/55d888a616be3c94d8e4073b4d1580541692997d -> trunk/55d888a616be3c94d8e4073b4d1580541692997d 2025-07-17T08:34:52.0758777Z * [new tag] trunk/55dae0bf7a4b501be91d0e022e675337dfe288ff -> trunk/55dae0bf7a4b501be91d0e022e675337dfe288ff 2025-07-17T08:34:52.0761043Z * [new tag] trunk/55ef7b15e0b2de903bfc26adfb0788ecfbcb4ed4 -> trunk/55ef7b15e0b2de903bfc26adfb0788ecfbcb4ed4 2025-07-17T08:34:52.0763321Z * [new tag] trunk/5606c516fd87e5c3594177e4ca64c3cac7fdafd5 -> trunk/5606c516fd87e5c3594177e4ca64c3cac7fdafd5 2025-07-17T08:34:52.0765424Z * [new tag] trunk/5633283574c458bd6a3cbb6a0a890f0cb9c8b2b5 -> trunk/5633283574c458bd6a3cbb6a0a890f0cb9c8b2b5 2025-07-17T08:34:52.0767769Z * [new tag] trunk/563fd95563c5edd732ae260b3bd3d0c38822ab57 -> trunk/563fd95563c5edd732ae260b3bd3d0c38822ab57 2025-07-17T08:34:52.0770028Z * [new tag] trunk/565fd079099d33a81c11d7b36581f09441ba6efa -> trunk/565fd079099d33a81c11d7b36581f09441ba6efa 2025-07-17T08:34:52.0772137Z * [new tag] trunk/568ca89bac9a80a66d664593a26ae69ac604796e -> trunk/568ca89bac9a80a66d664593a26ae69ac604796e 2025-07-17T08:34:52.0774401Z * [new tag] trunk/5692cbb818f1af9fcce9234e202d232afa5f3fa3 -> trunk/5692cbb818f1af9fcce9234e202d232afa5f3fa3 2025-07-17T08:34:52.0776718Z * [new tag] trunk/56b03df6ac5b4185a2b7b92f253565500a5b51ca -> trunk/56b03df6ac5b4185a2b7b92f253565500a5b51ca 2025-07-17T08:34:52.0778991Z * [new tag] trunk/56b3bf0c74f5837fd4fa6293bc515d353cb10295 -> trunk/56b3bf0c74f5837fd4fa6293bc515d353cb10295 2025-07-17T08:34:52.0781239Z * [new tag] trunk/56c69bedcc7e2211e5a3d6249e51b1674be5d10e -> trunk/56c69bedcc7e2211e5a3d6249e51b1674be5d10e 2025-07-17T08:34:52.0783505Z * [new tag] trunk/5763ec5f8d11df5eea962bedc74563394c0e273f -> trunk/5763ec5f8d11df5eea962bedc74563394c0e273f 2025-07-17T08:34:52.0785904Z * [new tag] trunk/577baa411675ed241c7d5cf79f25c13e29dac583 -> trunk/577baa411675ed241c7d5cf79f25c13e29dac583 2025-07-17T08:34:52.0788190Z * [new tag] trunk/57e4d7b5ccb8e35b434b62cc1f9aafb7d4d82b55 -> trunk/57e4d7b5ccb8e35b434b62cc1f9aafb7d4d82b55 2025-07-17T08:34:52.0790427Z * [new tag] trunk/584a0510b30b2472e54197d6b67b6f5f5e8ac807 -> trunk/584a0510b30b2472e54197d6b67b6f5f5e8ac807 2025-07-17T08:34:52.0792689Z * [new tag] trunk/588b5fb94bc6d2195ba5a4287b7feebe736da7db -> trunk/588b5fb94bc6d2195ba5a4287b7feebe736da7db 2025-07-17T08:34:52.0795051Z * [new tag] trunk/58e5d20c57e6baafd82b3b49c0e52931f5ae5d51 -> trunk/58e5d20c57e6baafd82b3b49c0e52931f5ae5d51 2025-07-17T08:34:52.0797347Z * [new tag] trunk/590607c5992d200b86361c5b68d53c93d8099193 -> trunk/590607c5992d200b86361c5b68d53c93d8099193 2025-07-17T08:34:52.0799643Z * [new tag] trunk/590fe4d2d7565f2045ef1ad4f4aad1f3b3de7aa3 -> trunk/590fe4d2d7565f2045ef1ad4f4aad1f3b3de7aa3 2025-07-17T08:34:52.0801928Z * [new tag] trunk/5951fcd50acc51bb91beae8488758f35219da849 -> trunk/5951fcd50acc51bb91beae8488758f35219da849 2025-07-17T08:34:52.0804260Z * [new tag] trunk/596b418391aa5d520e69310250c1f86c6c0a1107 -> trunk/596b418391aa5d520e69310250c1f86c6c0a1107 2025-07-17T08:34:52.0806510Z * [new tag] trunk/59c3cac4547aafd2f718b7c64053098cc5886878 -> trunk/59c3cac4547aafd2f718b7c64053098cc5886878 2025-07-17T08:34:52.0808964Z * [new tag] trunk/59eb61b2d1e4b64debbefa036acd0d8c7d55f0a3 -> trunk/59eb61b2d1e4b64debbefa036acd0d8c7d55f0a3 2025-07-17T08:34:52.0811376Z * [new tag] trunk/59f9b25f3cfc635053843372ea29ff4bf754da3f -> trunk/59f9b25f3cfc635053843372ea29ff4bf754da3f 2025-07-17T08:34:52.0813818Z * [new tag] trunk/5a2db5152d23f76dbb45d20008d9af68e761e8d1 -> trunk/5a2db5152d23f76dbb45d20008d9af68e761e8d1 2025-07-17T08:34:52.0816242Z * [new tag] trunk/5a533f74a160f6a570c2b9770ffdd89874ba59e2 -> trunk/5a533f74a160f6a570c2b9770ffdd89874ba59e2 2025-07-17T08:34:52.0818357Z * [new tag] trunk/5a54db14e3843cfa87fd8d27487dbf2f2dfb6c47 -> trunk/5a54db14e3843cfa87fd8d27487dbf2f2dfb6c47 2025-07-17T08:34:52.0820692Z * [new tag] trunk/5a5a05a6a3be376130848e235df73b752eef0230 -> trunk/5a5a05a6a3be376130848e235df73b752eef0230 2025-07-17T08:34:52.0823063Z * [new tag] trunk/5ab257c74c8e3ffe99380790de2134ba8013555e -> trunk/5ab257c74c8e3ffe99380790de2134ba8013555e 2025-07-17T08:34:52.0825359Z * [new tag] trunk/5ab6a3fb6fd37c542060c606edd4b95c7e3cae82 -> trunk/5ab6a3fb6fd37c542060c606edd4b95c7e3cae82 2025-07-17T08:34:52.0827660Z * [new tag] trunk/5ad2bee2c8a7defd2580bb138145a49c37146fcc -> trunk/5ad2bee2c8a7defd2580bb138145a49c37146fcc 2025-07-17T08:34:52.0829930Z * [new tag] trunk/5aee022d8b2bc9d31ddaf877315ffb8ad9d62985 -> trunk/5aee022d8b2bc9d31ddaf877315ffb8ad9d62985 2025-07-17T08:34:52.0832038Z * [new tag] trunk/5b10b0a96f9abf8c2751db324f0773aa433ec783 -> trunk/5b10b0a96f9abf8c2751db324f0773aa433ec783 2025-07-17T08:34:52.0834351Z * [new tag] trunk/5b4e0255d7ed756c312fb74ffcf17fe62c903bd7 -> trunk/5b4e0255d7ed756c312fb74ffcf17fe62c903bd7 2025-07-17T08:34:52.0836936Z * [new tag] trunk/5b656289064311e8bb08e1148db7d9b41f9d7ac8 -> trunk/5b656289064311e8bb08e1148db7d9b41f9d7ac8 2025-07-17T08:34:52.0840205Z * [new tag] trunk/5b9db4335e61c1c903cb0769282cbea588e49036 -> trunk/5b9db4335e61c1c903cb0769282cbea588e49036 2025-07-17T08:34:52.0842578Z * [new tag] trunk/5bd7804be2186f1756fba75ecc900d3b19016ad5 -> trunk/5bd7804be2186f1756fba75ecc900d3b19016ad5 2025-07-17T08:34:52.0844881Z * [new tag] trunk/5c79a55e7e58c6382c7ce02da1cd07d358239d94 -> trunk/5c79a55e7e58c6382c7ce02da1cd07d358239d94 2025-07-17T08:34:52.0847221Z * [new tag] trunk/5c7e1d39ab81647300d60e7cc22b24acf2641457 -> trunk/5c7e1d39ab81647300d60e7cc22b24acf2641457 2025-07-17T08:34:52.0849526Z * [new tag] trunk/5cc4e856fda4accb2e9291527693a91dc2a18d89 -> trunk/5cc4e856fda4accb2e9291527693a91dc2a18d89 2025-07-17T08:34:52.0851924Z * [new tag] trunk/5cfe4377d6f3b5845ba1f67122e5fbda8d642ecd -> trunk/5cfe4377d6f3b5845ba1f67122e5fbda8d642ecd 2025-07-17T08:34:52.0854173Z * [new tag] trunk/5d5a5b3501dfb0759ed36d0a88b65cdcd87c1e27 -> trunk/5d5a5b3501dfb0759ed36d0a88b65cdcd87c1e27 2025-07-17T08:34:52.0856470Z * [new tag] trunk/5d8d126249f83a9581f6b086f0891753bbb7175e -> trunk/5d8d126249f83a9581f6b086f0891753bbb7175e 2025-07-17T08:34:52.0859225Z * [new tag] trunk/5db9a2b54ae69917753be5b2eef0e15ad3cfd19b -> trunk/5db9a2b54ae69917753be5b2eef0e15ad3cfd19b 2025-07-17T08:34:52.0861617Z * [new tag] trunk/5dc75f72d4ede12067ddca459e02c8ee8a95e02c -> trunk/5dc75f72d4ede12067ddca459e02c8ee8a95e02c 2025-07-17T08:34:52.0863921Z * [new tag] trunk/5dd07c70e53a86b73f49711b8186d86dc4f1b32a -> trunk/5dd07c70e53a86b73f49711b8186d86dc4f1b32a 2025-07-17T08:34:52.0866119Z * [new tag] trunk/5dd9652389ed7959a842323e4ce063f553710e47 -> trunk/5dd9652389ed7959a842323e4ce063f553710e47 2025-07-17T08:34:52.0868533Z * [new tag] trunk/5df3bf13ec4e436abefe9d3822230727c04c2ab7 -> trunk/5df3bf13ec4e436abefe9d3822230727c04c2ab7 2025-07-17T08:34:52.0870806Z * [new tag] trunk/5dfd8a9c7a464bb42e81b8594eefd2fa865e5423 -> trunk/5dfd8a9c7a464bb42e81b8594eefd2fa865e5423 2025-07-17T08:34:52.0873167Z * [new tag] trunk/5dfe1787b5ed372fe1e9f0d73fab4e53a61bd17a -> trunk/5dfe1787b5ed372fe1e9f0d73fab4e53a61bd17a 2025-07-17T08:34:52.0875382Z * [new tag] trunk/5e18bc333144473f1f10bc8a5ba05dba7950fb8a -> trunk/5e18bc333144473f1f10bc8a5ba05dba7950fb8a 2025-07-17T08:34:52.0877980Z * [new tag] trunk/5e636d664ae0ccfdc1f0c5b060eb50fabf56e91b -> trunk/5e636d664ae0ccfdc1f0c5b060eb50fabf56e91b 2025-07-17T08:34:52.0880143Z * [new tag] trunk/5e93abe3c0106cf6fc694a38433909f20ee6a92c -> trunk/5e93abe3c0106cf6fc694a38433909f20ee6a92c 2025-07-17T08:34:52.0882451Z * [new tag] trunk/5eb5c3700bf51015e6a46f9f20c0e0e20cf55ab6 -> trunk/5eb5c3700bf51015e6a46f9f20c0e0e20cf55ab6 2025-07-17T08:34:52.0884705Z * [new tag] trunk/5f1225ef487a4c3954499a8ae5b5a55751e3ef2b -> trunk/5f1225ef487a4c3954499a8ae5b5a55751e3ef2b 2025-07-17T08:34:52.0887033Z * [new tag] trunk/5fb07acbc32874a932cd26087cf752b2f4cc72df -> trunk/5fb07acbc32874a932cd26087cf752b2f4cc72df 2025-07-17T08:34:52.0889909Z * [new tag] trunk/5fbaa041e75af11c7ff892e544707820232bd750 -> trunk/5fbaa041e75af11c7ff892e544707820232bd750 2025-07-17T08:34:52.0892187Z * [new tag] trunk/602044068359a3be08b2030a1bb1efb4dc107767 -> trunk/602044068359a3be08b2030a1bb1efb4dc107767 2025-07-17T08:34:52.0894512Z * [new tag] trunk/603a54a9b33e1aabe1407721d7935b881a160968 -> trunk/603a54a9b33e1aabe1407721d7935b881a160968 2025-07-17T08:34:52.0896841Z * [new tag] trunk/606d73bde495b055984206986fb7dd8918570e89 -> trunk/606d73bde495b055984206986fb7dd8918570e89 2025-07-17T08:34:52.0899256Z * [new tag] trunk/6098209bfffa453ec2cea08b0c9991b3b7e2bfeb -> trunk/6098209bfffa453ec2cea08b0c9991b3b7e2bfeb 2025-07-17T08:34:52.0901485Z * [new tag] trunk/60abb0d3273749cb2a7d583c7c2863bd2819e87e -> trunk/60abb0d3273749cb2a7d583c7c2863bd2819e87e 2025-07-17T08:34:52.0903719Z * [new tag] trunk/60b41de0ca0e5240eccf15c26a2b55b4a34d6f47 -> trunk/60b41de0ca0e5240eccf15c26a2b55b4a34d6f47 2025-07-17T08:34:52.0905958Z * [new tag] trunk/60e66d11ab3d62273cafa7f6b104db7e20f75dfd -> trunk/60e66d11ab3d62273cafa7f6b104db7e20f75dfd 2025-07-17T08:34:52.0908228Z * [new tag] trunk/61712e6f2ba58cce354a742d918934ec7293ee43 -> trunk/61712e6f2ba58cce354a742d918934ec7293ee43 2025-07-17T08:34:52.0910556Z * [new tag] trunk/617e3f69f8479197be57a28cc31e001c3feec407 -> trunk/617e3f69f8479197be57a28cc31e001c3feec407 2025-07-17T08:34:52.0912774Z * [new tag] trunk/61a7b09ef39907b6c4b47445f965c4c77cc6c2ed -> trunk/61a7b09ef39907b6c4b47445f965c4c77cc6c2ed 2025-07-17T08:34:52.0917537Z * [new tag] trunk/61b271e0f3f93209325dea9dccb1e97e7bc16b41 -> trunk/61b271e0f3f93209325dea9dccb1e97e7bc16b41 2025-07-17T08:34:52.0920108Z * [new tag] trunk/61e13782ddddf9e957c984ef11d7bb7643b871e7 -> trunk/61e13782ddddf9e957c984ef11d7bb7643b871e7 2025-07-17T08:34:52.0922511Z * [new tag] trunk/61eaaa21a42398941ea3fb01585a0926e9544831 -> trunk/61eaaa21a42398941ea3fb01585a0926e9544831 2025-07-17T08:34:52.0924779Z * [new tag] trunk/61f6aa36b9f9a168bd3c710c449f0a78c0301972 -> trunk/61f6aa36b9f9a168bd3c710c449f0a78c0301972 2025-07-17T08:34:52.0927021Z * [new tag] trunk/6200584193b770411b7f91880bbff6f746acfcb0 -> trunk/6200584193b770411b7f91880bbff6f746acfcb0 2025-07-17T08:34:52.0929304Z * [new tag] trunk/620415e018cd67e45d5c0a87964aff8751c9bf71 -> trunk/620415e018cd67e45d5c0a87964aff8751c9bf71 2025-07-17T08:34:52.0931580Z * [new tag] trunk/6215e90b7b9af8275c5dbfaa5fd58d7ec08b6764 -> trunk/6215e90b7b9af8275c5dbfaa5fd58d7ec08b6764 2025-07-17T08:34:52.0933814Z * [new tag] trunk/62272d5b24e7f505a02175de3c56ecc287557d2a -> trunk/62272d5b24e7f505a02175de3c56ecc287557d2a 2025-07-17T08:34:52.0936109Z * [new tag] trunk/627ba411366bcc15019c49756d3f22fd3914bd50 -> trunk/627ba411366bcc15019c49756d3f22fd3914bd50 2025-07-17T08:34:52.0938456Z * [new tag] trunk/62fa3f5aebfe763f5de27e687e4c81df7c7080e3 -> trunk/62fa3f5aebfe763f5de27e687e4c81df7c7080e3 2025-07-17T08:34:52.0940793Z * [new tag] trunk/6303cc41b71c02af9c855f3cc85b1146a97fc7a8 -> trunk/6303cc41b71c02af9c855f3cc85b1146a97fc7a8 2025-07-17T08:34:52.0943166Z * [new tag] trunk/63360e64da814de8ce271f1e4b6e2380a03b585e -> trunk/63360e64da814de8ce271f1e4b6e2380a03b585e 2025-07-17T08:34:52.0945295Z * [new tag] trunk/63a96eaeb84f5af6e83c7360c09f5540a44d19ca -> trunk/63a96eaeb84f5af6e83c7360c09f5540a44d19ca 2025-07-17T08:34:52.0947603Z * [new tag] trunk/63e87d6d05dc05ed7cc642c0cf8fd1cd2e3b93a1 -> trunk/63e87d6d05dc05ed7cc642c0cf8fd1cd2e3b93a1 2025-07-17T08:34:52.0949873Z * [new tag] trunk/6401d1d53d13cb2d564d48a30a8cf4f952ff773d -> trunk/6401d1d53d13cb2d564d48a30a8cf4f952ff773d 2025-07-17T08:34:52.0952225Z * [new tag] trunk/640703d95f210876f20f8f16c868442dad17b477 -> trunk/640703d95f210876f20f8f16c868442dad17b477 2025-07-17T08:34:52.0954647Z * [new tag] trunk/640f5a70905b01405f1ce397d25308d3e6a1d421 -> trunk/640f5a70905b01405f1ce397d25308d3e6a1d421 2025-07-17T08:34:52.0957118Z * [new tag] trunk/6442ae9256a6936bc8efe31fc5326d5c1c1f660b -> trunk/6442ae9256a6936bc8efe31fc5326d5c1c1f660b 2025-07-17T08:34:52.0959336Z * [new tag] trunk/64436c38c956b252d8c9f6f68e94fb46c34efdac -> trunk/64436c38c956b252d8c9f6f68e94fb46c34efdac 2025-07-17T08:34:52.0961574Z * [new tag] trunk/644cc58dfffe1b5bd15688495551b49462c163f6 -> trunk/644cc58dfffe1b5bd15688495551b49462c163f6 2025-07-17T08:34:52.0963932Z * [new tag] trunk/64bb6317a57d14c37339d86cded7c6b860c7d638 -> trunk/64bb6317a57d14c37339d86cded7c6b860c7d638 2025-07-17T08:34:52.0966145Z * [new tag] trunk/64f2ec77f869a7d495694519fb482e9ecaaa6da1 -> trunk/64f2ec77f869a7d495694519fb482e9ecaaa6da1 2025-07-17T08:34:52.0968547Z * [new tag] trunk/651b4a68f2a60d55d266e40776709247ef347d68 -> trunk/651b4a68f2a60d55d266e40776709247ef347d68 2025-07-17T08:34:52.0970814Z * [new tag] trunk/653c52fe52254e6783d75a73c3a8abbfd9cd2b3c -> trunk/653c52fe52254e6783d75a73c3a8abbfd9cd2b3c 2025-07-17T08:34:52.0973070Z * [new tag] trunk/655b3b14ffba4ae73e26a63b4289329e8d160a6f -> trunk/655b3b14ffba4ae73e26a63b4289329e8d160a6f 2025-07-17T08:34:52.0975275Z * [new tag] trunk/65b9c13cce43a7a8666ca1ef12b554a5dcfd2492 -> trunk/65b9c13cce43a7a8666ca1ef12b554a5dcfd2492 2025-07-17T08:34:52.0977549Z * [new tag] trunk/65fcca4f8c97de82d35d51ad9b790d10433e9b91 -> trunk/65fcca4f8c97de82d35d51ad9b790d10433e9b91 2025-07-17T08:34:52.0979820Z * [new tag] trunk/660695f11dd73efbe2694be0ed5826e4624f22e4 -> trunk/660695f11dd73efbe2694be0ed5826e4624f22e4 2025-07-17T08:34:52.0982125Z * [new tag] trunk/660dbea909b98f19dc0239a7ae565f8a52cfe846 -> trunk/660dbea909b98f19dc0239a7ae565f8a52cfe846 2025-07-17T08:34:52.0984446Z * [new tag] trunk/6629eaf0c658b523eed47a98c0a3e6b44fd89c6b -> trunk/6629eaf0c658b523eed47a98c0a3e6b44fd89c6b 2025-07-17T08:34:52.0986699Z * [new tag] trunk/662c1cfed2bf3262709f1150fcb11e96bc29e142 -> trunk/662c1cfed2bf3262709f1150fcb11e96bc29e142 2025-07-17T08:34:52.0988976Z * [new tag] trunk/670dab6c630552b32189911f22896ec453e55ab7 -> trunk/670dab6c630552b32189911f22896ec453e55ab7 2025-07-17T08:34:52.0991237Z * [new tag] trunk/671a9d175b7e15b4947c97f50b931dba1cc35a63 -> trunk/671a9d175b7e15b4947c97f50b931dba1cc35a63 2025-07-17T08:34:52.0993609Z * [new tag] trunk/672ac2ec86d8ffe080399363df96597437e51115 -> trunk/672ac2ec86d8ffe080399363df96597437e51115 2025-07-17T08:34:52.0996013Z * [new tag] trunk/67ee0c6725e8dd2d0372ff902775b7040631aaf7 -> trunk/67ee0c6725e8dd2d0372ff902775b7040631aaf7 2025-07-17T08:34:52.0998265Z * [new tag] trunk/67f8270516ef877aec85801e4a8d0b533687e938 -> trunk/67f8270516ef877aec85801e4a8d0b533687e938 2025-07-17T08:34:52.1000334Z * [new tag] trunk/6835ba1b3409d9764c6e704cdc0311d53ea0841a -> trunk/6835ba1b3409d9764c6e704cdc0311d53ea0841a 2025-07-17T08:34:52.1002775Z * [new tag] trunk/68996dc18306ac0320ac5e6979f1a7a7a226b3cc -> trunk/68996dc18306ac0320ac5e6979f1a7a7a226b3cc 2025-07-17T08:34:52.1004933Z * [new tag] trunk/68f36683f0c0dfe7befeba2b65dee30faf88f7cc -> trunk/68f36683f0c0dfe7befeba2b65dee30faf88f7cc 2025-07-17T08:34:52.1007198Z * [new tag] trunk/6918758f554790e1155fe9f3ee4120692347680a -> trunk/6918758f554790e1155fe9f3ee4120692347680a 2025-07-17T08:34:52.1009529Z * [new tag] trunk/693116f765ede6632d3dd8ef3ffe01cc40b960b9 -> trunk/693116f765ede6632d3dd8ef3ffe01cc40b960b9 2025-07-17T08:34:52.1011985Z * [new tag] trunk/694028f50269c9f34721db813e9ee1a8221c99b2 -> trunk/694028f50269c9f34721db813e9ee1a8221c99b2 2025-07-17T08:34:52.1014367Z * [new tag] trunk/6959b5febe090375e261e678902681533d2e2528 -> trunk/6959b5febe090375e261e678902681533d2e2528 2025-07-17T08:34:52.1016623Z * [new tag] trunk/69f2e09cc241176d6128a8f92f9490422c5b81b3 -> trunk/69f2e09cc241176d6128a8f92f9490422c5b81b3 2025-07-17T08:34:52.1018911Z * [new tag] trunk/6a3d00aa3b1e54feeb040891bfe2c59c36b9443c -> trunk/6a3d00aa3b1e54feeb040891bfe2c59c36b9443c 2025-07-17T08:34:52.1021181Z * [new tag] trunk/6abe450a6fda580714a276ed2dd695ab99074b93 -> trunk/6abe450a6fda580714a276ed2dd695ab99074b93 2025-07-17T08:34:52.1023418Z * [new tag] trunk/6b05842e4785f52ab89a8fa2a88f9e6f2316cbbd -> trunk/6b05842e4785f52ab89a8fa2a88f9e6f2316cbbd 2025-07-17T08:34:52.1025796Z * [new tag] trunk/6b1211df294e57d59c1e1717b1fedc671ec5bd5a -> trunk/6b1211df294e57d59c1e1717b1fedc671ec5bd5a 2025-07-17T08:34:52.1027986Z * [new tag] trunk/6b2bef10afae4acb18f230a496392b673c954ce7 -> trunk/6b2bef10afae4acb18f230a496392b673c954ce7 2025-07-17T08:34:52.1030293Z * [new tag] trunk/6b3eef6d316ce33ef0f0416c2f3f6a827ac4fe30 -> trunk/6b3eef6d316ce33ef0f0416c2f3f6a827ac4fe30 2025-07-17T08:34:52.1032576Z * [new tag] trunk/6b45af38a5991134f45d9750b79c767688ba3761 -> trunk/6b45af38a5991134f45d9750b79c767688ba3761 2025-07-17T08:34:52.1035019Z * [new tag] trunk/6b7767fc8d759017b5fc2a7002051cf5c05615a9 -> trunk/6b7767fc8d759017b5fc2a7002051cf5c05615a9 2025-07-17T08:34:52.1037345Z * [new tag] trunk/6b84cb29f97957032c5aa91b4e2a6bfa166db04d -> trunk/6b84cb29f97957032c5aa91b4e2a6bfa166db04d 2025-07-17T08:34:52.1039557Z * [new tag] trunk/6bc263809de610e2fcdf594c122a0f99c6c7d96f -> trunk/6bc263809de610e2fcdf594c122a0f99c6c7d96f 2025-07-17T08:34:52.1041842Z * [new tag] trunk/6c008e2fb5a94342c5c4047d034c61bb067fa2b7 -> trunk/6c008e2fb5a94342c5c4047d034c61bb067fa2b7 2025-07-17T08:34:52.1044111Z * [new tag] trunk/6c05f2fca049344fbbb55eed7a6c866f34c8a477 -> trunk/6c05f2fca049344fbbb55eed7a6c866f34c8a477 2025-07-17T08:34:52.1046369Z * [new tag] trunk/6c0b42fd2f2c070f4c7ef9a3914698726f61ef3a -> trunk/6c0b42fd2f2c070f4c7ef9a3914698726f61ef3a 2025-07-17T08:34:52.1048702Z * [new tag] trunk/6c24c6633a00608d8c20805a77fae55f6b8fe9a6 -> trunk/6c24c6633a00608d8c20805a77fae55f6b8fe9a6 2025-07-17T08:34:52.1051141Z * [new tag] trunk/6c42afe1964496e1b02be1442470f254e35eb199 -> trunk/6c42afe1964496e1b02be1442470f254e35eb199 2025-07-17T08:34:52.1053352Z * [new tag] trunk/6c5227ba00a2904365af566c24b4681cd01a041c -> trunk/6c5227ba00a2904365af566c24b4681cd01a041c 2025-07-17T08:34:52.1056090Z * [new tag] trunk/6c795306378c47341d58109da03371bba2bec46e -> trunk/6c795306378c47341d58109da03371bba2bec46e 2025-07-17T08:34:52.1058417Z * [new tag] trunk/6cc490d40b16b3be7ecc8943fe1bc594246caab3 -> trunk/6cc490d40b16b3be7ecc8943fe1bc594246caab3 2025-07-17T08:34:52.1060690Z * [new tag] trunk/6d02321472ee0761092166dd273eb3ec386cf0c0 -> trunk/6d02321472ee0761092166dd273eb3ec386cf0c0 2025-07-17T08:34:52.1063117Z * [new tag] trunk/6d2155db498375317e2101068a220439dba80939 -> trunk/6d2155db498375317e2101068a220439dba80939 2025-07-17T08:34:52.1065110Z * [new tag] trunk/6d3a4356f61b28a14abd95f641e2615deb186365 -> trunk/6d3a4356f61b28a14abd95f641e2615deb186365 2025-07-17T08:34:52.1067386Z * [new tag] trunk/6d5c789ad5f5d162426a5a10c84813ec923dd8ee -> trunk/6d5c789ad5f5d162426a5a10c84813ec923dd8ee 2025-07-17T08:34:52.1069719Z * [new tag] trunk/6dc2b22269f82a1feec06b5e3e79b184704cc0ba -> trunk/6dc2b22269f82a1feec06b5e3e79b184704cc0ba 2025-07-17T08:34:52.1071965Z * [new tag] trunk/6de41ce0f899604c3f8b33e1f8d37eb89b3a963e -> trunk/6de41ce0f899604c3f8b33e1f8d37eb89b3a963e 2025-07-17T08:34:52.1074330Z * [new tag] trunk/6defd5084e3fd330561095252c30ea50efa8e0ca -> trunk/6defd5084e3fd330561095252c30ea50efa8e0ca 2025-07-17T08:34:52.1076854Z * [new tag] trunk/6dfada220e4f46c28df3a6bdf57260ca0717c4bd -> trunk/6dfada220e4f46c28df3a6bdf57260ca0717c4bd 2025-07-17T08:34:52.1079101Z * [new tag] trunk/6e17315cd33892919dc6deb6a9ecb568dda6195e -> trunk/6e17315cd33892919dc6deb6a9ecb568dda6195e 2025-07-17T08:34:52.1081476Z * [new tag] trunk/6e185c53124e1b5a0fe391959060c1249178bcb6 -> trunk/6e185c53124e1b5a0fe391959060c1249178bcb6 2025-07-17T08:34:52.1083732Z * [new tag] trunk/6e2992a9984f2c3f6469564008c7e45869b84678 -> trunk/6e2992a9984f2c3f6469564008c7e45869b84678 2025-07-17T08:34:52.1086112Z * [new tag] trunk/6ea91f067256447cda6fae533f806c1f8baafbe2 -> trunk/6ea91f067256447cda6fae533f806c1f8baafbe2 2025-07-17T08:34:52.1088379Z * [new tag] trunk/6eb6f198e1a28e904ce00a132c43aa2536419664 -> trunk/6eb6f198e1a28e904ce00a132c43aa2536419664 2025-07-17T08:34:52.1090606Z * [new tag] trunk/6ebe9a4f47e9cd1c9ccd467bcdfdea9445fd98d6 -> trunk/6ebe9a4f47e9cd1c9ccd467bcdfdea9445fd98d6 2025-07-17T08:34:52.1092874Z * [new tag] trunk/6ed85bfe6a1704a87a492a3381414186eb439f5e -> trunk/6ed85bfe6a1704a87a492a3381414186eb439f5e 2025-07-17T08:34:52.1095197Z * [new tag] trunk/6ef70edd9a9acee76a607bdd22bac400c8e29434 -> trunk/6ef70edd9a9acee76a607bdd22bac400c8e29434 2025-07-17T08:34:52.1097475Z * [new tag] trunk/6f05d58f2b0bc4779fbed9db5b0f80af1c686dcd -> trunk/6f05d58f2b0bc4779fbed9db5b0f80af1c686dcd 2025-07-17T08:34:52.1099753Z * [new tag] trunk/6f23f53599629a47d6e097b2a027048658a142d4 -> trunk/6f23f53599629a47d6e097b2a027048658a142d4 2025-07-17T08:34:52.1102043Z * [new tag] trunk/6f60cfe9b1445271f8b83f58b4c4f12578d837d3 -> trunk/6f60cfe9b1445271f8b83f58b4c4f12578d837d3 2025-07-17T08:34:52.1104416Z * [new tag] trunk/6fb62931593fc10252e4994cd1a0595ebf8e990f -> trunk/6fb62931593fc10252e4994cd1a0595ebf8e990f 2025-07-17T08:34:52.1106761Z * [new tag] trunk/6fe7456aa1a2d025d1d06e15ba3896e6adba94b8 -> trunk/6fe7456aa1a2d025d1d06e15ba3896e6adba94b8 2025-07-17T08:34:52.1109099Z * [new tag] trunk/6ffa03ef9e18299f41da076f421abbd32e51e175 -> trunk/6ffa03ef9e18299f41da076f421abbd32e51e175 2025-07-17T08:34:52.1111351Z * [new tag] trunk/702a304b0785d2dfe3d9bd4eed535c38e969818d -> trunk/702a304b0785d2dfe3d9bd4eed535c38e969818d 2025-07-17T08:34:52.1113594Z * [new tag] trunk/706bc41c4c3caa2654a926382c2ef0ef4d0bc55e -> trunk/706bc41c4c3caa2654a926382c2ef0ef4d0bc55e 2025-07-17T08:34:52.1115949Z * [new tag] trunk/706e236b080cadd5b8a016e4bf972cdb0886d57b -> trunk/706e236b080cadd5b8a016e4bf972cdb0886d57b 2025-07-17T08:34:52.1118228Z * [new tag] trunk/7070ab318061d44f195a83b6dc11ef9299603dfe -> trunk/7070ab318061d44f195a83b6dc11ef9299603dfe 2025-07-17T08:34:52.1120496Z * [new tag] trunk/7081b8233a64c350c64e9f00c9b9d00e52020241 -> trunk/7081b8233a64c350c64e9f00c9b9d00e52020241 2025-07-17T08:34:52.1122882Z * [new tag] trunk/70b68caf58815419924ddeda231cbf6535181c53 -> trunk/70b68caf58815419924ddeda231cbf6535181c53 2025-07-17T08:34:52.1125384Z * [new tag] trunk/70bb34929aabc9c11d9ec9bbb53a4f9fc32e5e19 -> trunk/70bb34929aabc9c11d9ec9bbb53a4f9fc32e5e19 2025-07-17T08:34:52.1127433Z * [new tag] trunk/710b92cf3b577036d551708b351b4043817233f5 -> trunk/710b92cf3b577036d551708b351b4043817233f5 2025-07-17T08:34:52.1129856Z * [new tag] trunk/717a099d427d97a62c26fee58f9da9a0893d4233 -> trunk/717a099d427d97a62c26fee58f9da9a0893d4233 2025-07-17T08:34:52.1132119Z * [new tag] trunk/71a650ad565fe75ca752acbc8b40fb9ce677f40f -> trunk/71a650ad565fe75ca752acbc8b40fb9ce677f40f 2025-07-17T08:34:52.1134308Z * [new tag] trunk/720c2c46b181ad446bf970d70a27fd0ce149114d -> trunk/720c2c46b181ad446bf970d70a27fd0ce149114d 2025-07-17T08:34:52.1136617Z * [new tag] trunk/721d2580dbf9a4922adc1c7d1cc8237126d3cdd6 -> trunk/721d2580dbf9a4922adc1c7d1cc8237126d3cdd6 2025-07-17T08:34:52.1138935Z * [new tag] trunk/72453a66769294e93a183dfb5bdefa23de873d89 -> trunk/72453a66769294e93a183dfb5bdefa23de873d89 2025-07-17T08:34:52.1141279Z * [new tag] trunk/725c3272848c408d0fa2cba4de76affe90f793b5 -> trunk/725c3272848c408d0fa2cba4de76affe90f793b5 2025-07-17T08:34:52.1143537Z * [new tag] trunk/7275f280454f790414b24147a2ba7f94d0eabcf6 -> trunk/7275f280454f790414b24147a2ba7f94d0eabcf6 2025-07-17T08:34:52.1145792Z * [new tag] trunk/728cf6721e2996490922d0eacb23081953e45fc7 -> trunk/728cf6721e2996490922d0eacb23081953e45fc7 2025-07-17T08:34:52.1148058Z * [new tag] trunk/72c8751b61e0b2c7b88003a41ed737a65768063c -> trunk/72c8751b61e0b2c7b88003a41ed737a65768063c 2025-07-17T08:34:52.1150379Z * [new tag] trunk/731351bb4ac572fb47d46c70b7425c209c81570a -> trunk/731351bb4ac572fb47d46c70b7425c209c81570a 2025-07-17T08:34:52.1152607Z * [new tag] trunk/73220d52fd67b5f4f5b15e0e0433e09733c93f31 -> trunk/73220d52fd67b5f4f5b15e0e0433e09733c93f31 2025-07-17T08:34:52.1155328Z * [new tag] trunk/73772919d2db560c9cc8ed617c94c77450542a94 -> trunk/73772919d2db560c9cc8ed617c94c77450542a94 2025-07-17T08:34:52.1157694Z * [new tag] trunk/7381c777245a836f88e9622005c6ea16009a3a0d -> trunk/7381c777245a836f88e9622005c6ea16009a3a0d 2025-07-17T08:34:52.1160081Z * [new tag] trunk/7392470da4386e654f303eb526dfba7e7777b06b -> trunk/7392470da4386e654f303eb526dfba7e7777b06b 2025-07-17T08:34:52.1162353Z * [new tag] trunk/7444debaca7de16b2cf71d0cd82dc8998c3a73c4 -> trunk/7444debaca7de16b2cf71d0cd82dc8998c3a73c4 2025-07-17T08:34:52.1164668Z * [new tag] trunk/7485ef078f182af981b5f17d06602de68816492c -> trunk/7485ef078f182af981b5f17d06602de68816492c 2025-07-17T08:34:52.1166941Z * [new tag] trunk/749757ac1b66485459e751b0e58b5a7c64d5f561 -> trunk/749757ac1b66485459e751b0e58b5a7c64d5f561 2025-07-17T08:34:52.1169180Z * [new tag] trunk/74ebd8d14e0d612ed35557ab254a90facf73b029 -> trunk/74ebd8d14e0d612ed35557ab254a90facf73b029 2025-07-17T08:34:52.1171486Z * [new tag] trunk/7521cd91118c1c1a55b8f7146d7e22bf60fe06bf -> trunk/7521cd91118c1c1a55b8f7146d7e22bf60fe06bf 2025-07-17T08:34:52.1173788Z * [new tag] trunk/752f202ef3445d2e28e2abeb6feed56eee9da5a9 -> trunk/752f202ef3445d2e28e2abeb6feed56eee9da5a9 2025-07-17T08:34:52.1175985Z * [new tag] trunk/7531bd64911774fc2656f290601834d5d02d3925 -> trunk/7531bd64911774fc2656f290601834d5d02d3925 2025-07-17T08:34:52.1178311Z * [new tag] trunk/754699610b0abec2fe3f5a73269b1dd09a330445 -> trunk/754699610b0abec2fe3f5a73269b1dd09a330445 2025-07-17T08:34:52.1180594Z * [new tag] trunk/754c04aa062d8f3c0449aec4bbcaab00bfca4bf2 -> trunk/754c04aa062d8f3c0449aec4bbcaab00bfca4bf2 2025-07-17T08:34:52.1182838Z * [new tag] trunk/75824035d392595de3bd0b8588e63c2fc6d09139 -> trunk/75824035d392595de3bd0b8588e63c2fc6d09139 2025-07-17T08:34:52.1185160Z * [new tag] trunk/7597988f1b5a41c0b91d379e0ce51111fd7cc95a -> trunk/7597988f1b5a41c0b91d379e0ce51111fd7cc95a 2025-07-17T08:34:52.1187354Z * [new tag] trunk/7599bebead35dd21bad3d3e85f62f400ad2c3139 -> trunk/7599bebead35dd21bad3d3e85f62f400ad2c3139 2025-07-17T08:34:52.1189680Z * [new tag] trunk/75a7d9e86842b57f9001a0fa6fd716927614dbf9 -> trunk/75a7d9e86842b57f9001a0fa6fd716927614dbf9 2025-07-17T08:34:52.1191926Z * [new tag] trunk/75f258dd1feeb90b8084849155e049fb57ba3521 -> trunk/75f258dd1feeb90b8084849155e049fb57ba3521 2025-07-17T08:34:52.1194224Z * [new tag] trunk/75f3e5a88df60caef27fd9c9df3fd51161378fcc -> trunk/75f3e5a88df60caef27fd9c9df3fd51161378fcc 2025-07-17T08:34:52.1196838Z * [new tag] trunk/75f489d37f048ec8a9106163ca16201156f05499 -> trunk/75f489d37f048ec8a9106163ca16201156f05499 2025-07-17T08:34:52.1199166Z * [new tag] trunk/764c02b78b935ba61c46277cb7409e6419460058 -> trunk/764c02b78b935ba61c46277cb7409e6419460058 2025-07-17T08:34:52.1201433Z * [new tag] trunk/76644c9ff58285c95764ffd27d629df219aab140 -> trunk/76644c9ff58285c95764ffd27d629df219aab140 2025-07-17T08:34:52.1203825Z * [new tag] trunk/769d754ab2469813a3b790ec58c25c466099dd3d -> trunk/769d754ab2469813a3b790ec58c25c466099dd3d 2025-07-17T08:34:52.1206358Z * [new tag] trunk/76ca23c41c7edcf7b4c60ff6162eedcdf06ef359 -> trunk/76ca23c41c7edcf7b4c60ff6162eedcdf06ef359 2025-07-17T08:34:52.1208618Z * [new tag] trunk/76d07e919f66c24a55746bbc060c14b3df02ffa8 -> trunk/76d07e919f66c24a55746bbc060c14b3df02ffa8 2025-07-17T08:34:52.1210844Z * [new tag] trunk/76fe88fa56c8aac7377127fb5553de94e16e9070 -> trunk/76fe88fa56c8aac7377127fb5553de94e16e9070 2025-07-17T08:34:52.1213103Z * [new tag] trunk/770285522805745959221b902f23776a13d56df8 -> trunk/770285522805745959221b902f23776a13d56df8 2025-07-17T08:34:52.1215426Z * [new tag] trunk/7709ff55123dbdcaaa44004738fb767324f05dbd -> trunk/7709ff55123dbdcaaa44004738fb767324f05dbd 2025-07-17T08:34:52.1217836Z * [new tag] trunk/771be857043bf794cd219a9b925e308e31f12314 -> trunk/771be857043bf794cd219a9b925e308e31f12314 2025-07-17T08:34:52.1220596Z * [new tag] trunk/772d5904152abc9702bf49037e46ab6203b83f55 -> trunk/772d5904152abc9702bf49037e46ab6203b83f55 2025-07-17T08:34:52.1222852Z * [new tag] trunk/77518d1a13cc105637814bc157387478384dc897 -> trunk/77518d1a13cc105637814bc157387478384dc897 2025-07-17T08:34:52.1225099Z * [new tag] trunk/77676753ecabf6a6645bdd3abfe01939e5751e76 -> trunk/77676753ecabf6a6645bdd3abfe01939e5751e76 2025-07-17T08:34:52.1227482Z * [new tag] trunk/777eca9f16aeecd7c362a235cf25e6b8e6eda57f -> trunk/777eca9f16aeecd7c362a235cf25e6b8e6eda57f 2025-07-17T08:34:52.1229753Z * [new tag] trunk/77ac3a096532ee08211dae38c2a9336c970d1785 -> trunk/77ac3a096532ee08211dae38c2a9336c970d1785 2025-07-17T08:34:52.1232024Z * [new tag] trunk/77f884c2ec62df9df930ae86e9b8437364900346 -> trunk/77f884c2ec62df9df930ae86e9b8437364900346 2025-07-17T08:34:52.1234457Z * [new tag] trunk/783a4c1f5002bf1749833821c450360a417d40c7 -> trunk/783a4c1f5002bf1749833821c450360a417d40c7 2025-07-17T08:34:52.1236837Z * [new tag] trunk/78684e27ace1ebd51ccc93d4b6b8a6a3ec3b4c14 -> trunk/78684e27ace1ebd51ccc93d4b6b8a6a3ec3b4c14 2025-07-17T08:34:52.1239066Z * [new tag] trunk/78ee2ee90eed957aec3dc80423b108b16938a8ae -> trunk/78ee2ee90eed957aec3dc80423b108b16938a8ae 2025-07-17T08:34:52.1241310Z * [new tag] trunk/7918978653ff4f21c284b34809acf88784deb0de -> trunk/7918978653ff4f21c284b34809acf88784deb0de 2025-07-17T08:34:52.1243559Z * [new tag] trunk/794b95d54bb2c49d704169ddb777acfd8da8cfad -> trunk/794b95d54bb2c49d704169ddb777acfd8da8cfad 2025-07-17T08:34:52.1245810Z * [new tag] trunk/794ef6c9b8f173a0cbab5288efb2d68aef9c90b9 -> trunk/794ef6c9b8f173a0cbab5288efb2d68aef9c90b9 2025-07-17T08:34:52.1248124Z * [new tag] trunk/795a6a0affd349adfb4e3df298b604b74f27b44e -> trunk/795a6a0affd349adfb4e3df298b604b74f27b44e 2025-07-17T08:34:52.1250100Z * [new tag] trunk/7982b8c703e24a36c0e0a01b8cf22cd272369b28 -> trunk/7982b8c703e24a36c0e0a01b8cf22cd272369b28 2025-07-17T08:34:52.1252518Z * [new tag] trunk/7986c0dba6e1044d90b7f607f9cca15922339bb4 -> trunk/7986c0dba6e1044d90b7f607f9cca15922339bb4 2025-07-17T08:34:52.1254916Z * [new tag] trunk/799443605bffbd0d80fb3f2edbfe5517e0e2b4a3 -> trunk/799443605bffbd0d80fb3f2edbfe5517e0e2b4a3 2025-07-17T08:34:52.1257182Z * [new tag] trunk/7999735d23aeca844d4a7b23de6ac2370767099a -> trunk/7999735d23aeca844d4a7b23de6ac2370767099a 2025-07-17T08:34:52.1259503Z * [new tag] trunk/79ab84e9b8fe561a55931b2108af45993a670276 -> trunk/79ab84e9b8fe561a55931b2108af45993a670276 2025-07-17T08:34:52.1261781Z * [new tag] trunk/79aef141695f2daea4a9aeb0f385726c5794a242 -> trunk/79aef141695f2daea4a9aeb0f385726c5794a242 2025-07-17T08:34:52.1264048Z * [new tag] trunk/79bdafe5b61f6f073dcd276e135076ff3277a951 -> trunk/79bdafe5b61f6f073dcd276e135076ff3277a951 2025-07-17T08:34:52.1266356Z * [new tag] trunk/79d7c754ab8ae0e5c3a614521632d2cfbfa0fdba -> trunk/79d7c754ab8ae0e5c3a614521632d2cfbfa0fdba 2025-07-17T08:34:52.1268599Z * [new tag] trunk/7a03b0d2ca3946b770bc817571fc36bd5c608b38 -> trunk/7a03b0d2ca3946b770bc817571fc36bd5c608b38 2025-07-17T08:34:52.1270949Z * [new tag] trunk/7a08755c5f3630150c50d09e16c0abf9501dea1e -> trunk/7a08755c5f3630150c50d09e16c0abf9501dea1e 2025-07-17T08:34:52.1273510Z * [new tag] trunk/7a41f207943cc6c88160d7a531fc57bdbe149253 -> trunk/7a41f207943cc6c88160d7a531fc57bdbe149253 2025-07-17T08:34:52.1275990Z * [new tag] trunk/7a48cc699006879a6ec185ac5bd4eaad0c0b7e36 -> trunk/7a48cc699006879a6ec185ac5bd4eaad0c0b7e36 2025-07-17T08:34:52.1278276Z * [new tag] trunk/7a92b5119654c07d15f5c0818e6ae804b01e836c -> trunk/7a92b5119654c07d15f5c0818e6ae804b01e836c 2025-07-17T08:34:52.1280590Z * [new tag] trunk/7ae7c1414354ad225c4a36b3228e30a2baf94923 -> trunk/7ae7c1414354ad225c4a36b3228e30a2baf94923 2025-07-17T08:34:52.1282951Z * [new tag] trunk/7afb834f939eccbb3262e646f0922eed070074a7 -> trunk/7afb834f939eccbb3262e646f0922eed070074a7 2025-07-17T08:34:52.1285270Z * [new tag] trunk/7b0118884e0855e4918af53d44c3fbfc185a1ee9 -> trunk/7b0118884e0855e4918af53d44c3fbfc185a1ee9 2025-07-17T08:34:52.1287603Z * [new tag] trunk/7b392bac13a8007d1e0c6c789aa9bc515119faba -> trunk/7b392bac13a8007d1e0c6c789aa9bc515119faba 2025-07-17T08:34:52.1290088Z * [new tag] trunk/7b7cd56f5e2a965ae8a48c69441dd4f58bf68ceb -> trunk/7b7cd56f5e2a965ae8a48c69441dd4f58bf68ceb 2025-07-17T08:34:52.1292377Z * [new tag] trunk/7be862ab8f2b9040b34c58e1b77ce74c2512f062 -> trunk/7be862ab8f2b9040b34c58e1b77ce74c2512f062 2025-07-17T08:34:52.1294678Z * [new tag] trunk/7c1f62782875a4ac70b96aed0e6c9525b6e4eaf3 -> trunk/7c1f62782875a4ac70b96aed0e6c9525b6e4eaf3 2025-07-17T08:34:52.1296915Z * [new tag] trunk/7c51619e7fbd346e2299255c468cc43bd69425d7 -> trunk/7c51619e7fbd346e2299255c468cc43bd69425d7 2025-07-17T08:34:52.1299337Z * [new tag] trunk/7caf6c801ddfaf556a3ca191173b50002c4261f4 -> trunk/7caf6c801ddfaf556a3ca191173b50002c4261f4 2025-07-17T08:34:52.1301725Z * [new tag] trunk/7cda4017ddda554752e89069ae205be5e8388f59 -> trunk/7cda4017ddda554752e89069ae205be5e8388f59 2025-07-17T08:34:52.1304088Z * [new tag] trunk/7cf31b4a426f3791af30159cea420687f347cd7a -> trunk/7cf31b4a426f3791af30159cea420687f347cd7a 2025-07-17T08:34:52.1306463Z * [new tag] trunk/7cf38d2a0521c5ca292a720bce00617be749a0e7 -> trunk/7cf38d2a0521c5ca292a720bce00617be749a0e7 2025-07-17T08:34:52.1309009Z * [new tag] trunk/7cfd054075b0713642fab8dc4313fc1e5d992599 -> trunk/7cfd054075b0713642fab8dc4313fc1e5d992599 2025-07-17T08:34:52.1311162Z * [new tag] trunk/7d1b3f599d6968a3efe70b7401310e3127949f13 -> trunk/7d1b3f599d6968a3efe70b7401310e3127949f13 2025-07-17T08:34:52.1313647Z * [new tag] trunk/7d4228dbfd13d1ac8fac2c78c042dbb8314f042d -> trunk/7d4228dbfd13d1ac8fac2c78c042dbb8314f042d 2025-07-17T08:34:52.1316160Z * [new tag] trunk/7dcc77e422dcf97ce35991a138ab635a5cb88731 -> trunk/7dcc77e422dcf97ce35991a138ab635a5cb88731 2025-07-17T08:34:52.1318345Z * [new tag] trunk/7e433d5f423248914c5e9838d3ea145db7964923 -> trunk/7e433d5f423248914c5e9838d3ea145db7964923 2025-07-17T08:34:52.1320789Z * [new tag] trunk/7e4c097b0752ae79a8b5dd1de21a51aaafba2ef9 -> trunk/7e4c097b0752ae79a8b5dd1de21a51aaafba2ef9 2025-07-17T08:34:52.1323051Z * [new tag] trunk/7e54c02a35b905e758497b856a1953eb009ba836 -> trunk/7e54c02a35b905e758497b856a1953eb009ba836 2025-07-17T08:34:52.1326154Z * [new tag] trunk/7e83d5084588bdb51151f869b53c92c07bc9f544 -> trunk/7e83d5084588bdb51151f869b53c92c07bc9f544 2025-07-17T08:34:52.1328528Z * [new tag] trunk/7f0cddfb55d426b414cf7b4482c6a371618b349c -> trunk/7f0cddfb55d426b414cf7b4482c6a371618b349c 2025-07-17T08:34:52.1330699Z * [new tag] trunk/7f14b42adf70196d82340c59a9981ffcadf0c53c -> trunk/7f14b42adf70196d82340c59a9981ffcadf0c53c 2025-07-17T08:34:52.1332917Z * [new tag] trunk/7f6e7103a3aefd2d96b3cc6702be931a66b44977 -> trunk/7f6e7103a3aefd2d96b3cc6702be931a66b44977 2025-07-17T08:34:52.1335218Z * [new tag] trunk/7f9fc7e67ce9853a1bb4b16c901c708f78c1c5cd -> trunk/7f9fc7e67ce9853a1bb4b16c901c708f78c1c5cd 2025-07-17T08:34:52.1337477Z * [new tag] trunk/7fcad0231cb0ae9ba6a0004098b5ffd82c62858f -> trunk/7fcad0231cb0ae9ba6a0004098b5ffd82c62858f 2025-07-17T08:34:52.1339778Z * [new tag] trunk/805297981ae4e2aa08d133161477a477f5da274f -> trunk/805297981ae4e2aa08d133161477a477f5da274f 2025-07-17T08:34:52.1342046Z * [new tag] trunk/8088958793841191eba1faf98849904ee769bbfd -> trunk/8088958793841191eba1faf98849904ee769bbfd 2025-07-17T08:34:52.1344339Z * [new tag] trunk/80bcaa4195944d0f00f16abcbb702bde4d582d14 -> trunk/80bcaa4195944d0f00f16abcbb702bde4d582d14 2025-07-17T08:34:52.1346802Z * [new tag] trunk/80d89974c173c9d7a2c59fe9dfce2fe1301e2833 -> trunk/80d89974c173c9d7a2c59fe9dfce2fe1301e2833 2025-07-17T08:34:52.1349103Z * [new tag] trunk/8134684d44f3f9d09e775f3adc513bdef72245c9 -> trunk/8134684d44f3f9d09e775f3adc513bdef72245c9 2025-07-17T08:34:52.1351394Z * [new tag] trunk/8142a0286016e63a0e91b5667e1fb1a5e868ffd7 -> trunk/8142a0286016e63a0e91b5667e1fb1a5e868ffd7 2025-07-17T08:34:52.1353781Z * [new tag] trunk/8147c4a9044913e62ba3aab7f1eaaafead4f5930 -> trunk/8147c4a9044913e62ba3aab7f1eaaafead4f5930 2025-07-17T08:34:52.1356237Z * [new tag] trunk/8153340d1073c190ecf221e51abbd983286782a8 -> trunk/8153340d1073c190ecf221e51abbd983286782a8 2025-07-17T08:34:52.1359847Z * [new tag] trunk/815545f2dd6ade563cb1263f8bb7813f355edb2e -> trunk/815545f2dd6ade563cb1263f8bb7813f355edb2e 2025-07-17T08:34:52.1360799Z * [new tag] trunk/81759afed4fa4fa3a03f9ff671abe270679cc006 -> trunk/81759afed4fa4fa3a03f9ff671abe270679cc006 2025-07-17T08:34:52.1363126Z * [new tag] trunk/8186af5a26bc6a628f283c4791c8d68ef6fa1336 -> trunk/8186af5a26bc6a628f283c4791c8d68ef6fa1336 2025-07-17T08:34:52.1365426Z * [new tag] trunk/81b0b308cafae44339b8a123e8b1bec3bf9eb8ad -> trunk/81b0b308cafae44339b8a123e8b1bec3bf9eb8ad 2025-07-17T08:34:52.1367705Z * [new tag] trunk/81bf278537c081279a574157e1359ec7f0b65daf -> trunk/81bf278537c081279a574157e1359ec7f0b65daf 2025-07-17T08:34:52.1369916Z * [new tag] trunk/81c7445eb97f910ed89da9b11140d914651a1675 -> trunk/81c7445eb97f910ed89da9b11140d914651a1675 2025-07-17T08:34:52.1373064Z * [new tag] trunk/82672206b724a767dcb8c2541074449100de701b -> trunk/82672206b724a767dcb8c2541074449100de701b 2025-07-17T08:34:52.1374310Z * [new tag] trunk/826f12b829070e3d5bfd050f001b61aaf78e5447 -> trunk/826f12b829070e3d5bfd050f001b61aaf78e5447 2025-07-17T08:34:52.1376617Z * [new tag] trunk/82765dad16d7cda1ea382497dbf8c1ab2d07bfdd -> trunk/82765dad16d7cda1ea382497dbf8c1ab2d07bfdd 2025-07-17T08:34:52.1378943Z * [new tag] trunk/82a1ee1135b054d371d10081883b848ac7b7419f -> trunk/82a1ee1135b054d371d10081883b848ac7b7419f 2025-07-17T08:34:52.1381248Z * [new tag] trunk/82e6475d920991ef3be5d5637a72bf49313cc604 -> trunk/82e6475d920991ef3be5d5637a72bf49313cc604 2025-07-17T08:34:52.1383543Z * [new tag] trunk/82eefaedd98b63de8a87e34275af781f8eb177e1 -> trunk/82eefaedd98b63de8a87e34275af781f8eb177e1 2025-07-17T08:34:52.1385749Z * [new tag] trunk/82fb904140d258604a185154287b9c6fde047fc5 -> trunk/82fb904140d258604a185154287b9c6fde047fc5 2025-07-17T08:34:52.1387994Z * [new tag] trunk/830a335a7da5fec00395d440ba568749cb4e2e9e -> trunk/830a335a7da5fec00395d440ba568749cb4e2e9e 2025-07-17T08:34:52.1390230Z * [new tag] trunk/831c9010c7d7460fd78bc59b3fa28a65a0339e5d -> trunk/831c9010c7d7460fd78bc59b3fa28a65a0339e5d 2025-07-17T08:34:52.1392537Z * [new tag] trunk/8347268edcf3cc4f638474d860450bb0cd7a9908 -> trunk/8347268edcf3cc4f638474d860450bb0cd7a9908 2025-07-17T08:34:52.1394875Z * [new tag] trunk/836bb1941b593063f002394c1de3ec382c2ce50e -> trunk/836bb1941b593063f002394c1de3ec382c2ce50e 2025-07-17T08:34:52.1397258Z * [new tag] trunk/83700b448874cac8e89356ab06292e4289577269 -> trunk/83700b448874cac8e89356ab06292e4289577269 2025-07-17T08:34:52.1399466Z * [new tag] trunk/8372d0986a495d41f41c5d7684f1428ff847c699 -> trunk/8372d0986a495d41f41c5d7684f1428ff847c699 2025-07-17T08:34:52.1402112Z * [new tag] trunk/838798425731e6c90df4a885e90ec03f18eea10b -> trunk/838798425731e6c90df4a885e90ec03f18eea10b 2025-07-17T08:34:52.1404216Z * [new tag] trunk/83d22256f84232c5440b25a08459c649a32b9a4f -> trunk/83d22256f84232c5440b25a08459c649a32b9a4f 2025-07-17T08:34:52.1406526Z * [new tag] trunk/84085229765698166f07c9220d5544023ab80d47 -> trunk/84085229765698166f07c9220d5544023ab80d47 2025-07-17T08:34:52.1408759Z * [new tag] trunk/8485f1950710d383fd8d0d772fc978933c6fc175 -> trunk/8485f1950710d383fd8d0d772fc978933c6fc175 2025-07-17T08:34:52.1411028Z * [new tag] trunk/84b77ec1280a56bece8408a2e0e179b9eb95e1e2 -> trunk/84b77ec1280a56bece8408a2e0e179b9eb95e1e2 2025-07-17T08:34:52.1413291Z * [new tag] trunk/84c14361c28e1b553ea405ef5020cf0a90468850 -> trunk/84c14361c28e1b553ea405ef5020cf0a90468850 2025-07-17T08:34:52.1415554Z * [new tag] trunk/84c588e5eada9e7921608065edc444a15c22cb1c -> trunk/84c588e5eada9e7921608065edc444a15c22cb1c 2025-07-17T08:34:52.1417953Z * [new tag] trunk/84dec060b79078e68dcb6be1e7f308dad05d00e2 -> trunk/84dec060b79078e68dcb6be1e7f308dad05d00e2 2025-07-17T08:34:52.1420116Z * [new tag] trunk/85111cd165f108ffabb4a90083d59d7a867ebd9f -> trunk/85111cd165f108ffabb4a90083d59d7a867ebd9f 2025-07-17T08:34:52.1422421Z * [new tag] trunk/851a6fa82df251fbc368f0284d941ce78f68e7b1 -> trunk/851a6fa82df251fbc368f0284d941ce78f68e7b1 2025-07-17T08:34:52.1424662Z * [new tag] trunk/85320336799e38411d15c0e159b41248cda01218 -> trunk/85320336799e38411d15c0e159b41248cda01218 2025-07-17T08:34:52.1427147Z * [new tag] trunk/8554c8007ddaa8029e7e01bb1af12f358bf597c2 -> trunk/8554c8007ddaa8029e7e01bb1af12f358bf597c2 2025-07-17T08:34:52.1429448Z * [new tag] trunk/85857181ebca86e9c709e9922a9d9ef41a9c4ef9 -> trunk/85857181ebca86e9c709e9922a9d9ef41a9c4ef9 2025-07-17T08:34:52.1431977Z * [new tag] trunk/85df746892d9b0e87e7a5dfa78ef81a84aec6de0 -> trunk/85df746892d9b0e87e7a5dfa78ef81a84aec6de0 2025-07-17T08:34:52.1434070Z * [new tag] trunk/86251eff4069c468efebdb481dd18fe8d42856f0 -> trunk/86251eff4069c468efebdb481dd18fe8d42856f0 2025-07-17T08:34:52.1436558Z * [new tag] trunk/863327ae496471654344e1e04ccaa713a44a135d -> trunk/863327ae496471654344e1e04ccaa713a44a135d 2025-07-17T08:34:52.1438891Z * [new tag] trunk/86670b39fa3df63a652a9a06b59b73f92d70c392 -> trunk/86670b39fa3df63a652a9a06b59b73f92d70c392 2025-07-17T08:34:52.1441527Z * [new tag] trunk/86996c15dc4294c400c3a552d1a9e8e90aa6c7f6 -> trunk/86996c15dc4294c400c3a552d1a9e8e90aa6c7f6 2025-07-17T08:34:52.1443829Z * [new tag] trunk/86ced144534c2c78460a2d0f1d9ed2ba048f2514 -> trunk/86ced144534c2c78460a2d0f1d9ed2ba048f2514 2025-07-17T08:34:52.1446098Z * [new tag] trunk/86d8af6a6cc648134289de89d393d0dce5b3a5f4 -> trunk/86d8af6a6cc648134289de89d393d0dce5b3a5f4 2025-07-17T08:34:52.1448429Z * [new tag] trunk/86eaf452c330bd871262b7590ff8bd1bf432e2d7 -> trunk/86eaf452c330bd871262b7590ff8bd1bf432e2d7 2025-07-17T08:34:52.1450721Z * [new tag] trunk/87d615efab194482a00d241e4d9aebc513635cf5 -> trunk/87d615efab194482a00d241e4d9aebc513635cf5 2025-07-17T08:34:52.1452934Z * [new tag] trunk/8823138e47a3200c313f6bf2d21eb689d8150f39 -> trunk/8823138e47a3200c313f6bf2d21eb689d8150f39 2025-07-17T08:34:52.1455232Z * [new tag] trunk/8892b782a85b9de9f51334e65e009b3d59899b02 -> trunk/8892b782a85b9de9f51334e65e009b3d59899b02 2025-07-17T08:34:52.1457571Z * [new tag] trunk/88b9c285e0eb68134c5f834ceccdb99bc1d255d0 -> trunk/88b9c285e0eb68134c5f834ceccdb99bc1d255d0 2025-07-17T08:34:52.1459836Z * [new tag] trunk/88c6199db09372b6e2d55a5349ab545527842727 -> trunk/88c6199db09372b6e2d55a5349ab545527842727 2025-07-17T08:34:52.1462078Z * [new tag] trunk/88cd9f34b00c22e4edefa223490fcdd6c4ff272a -> trunk/88cd9f34b00c22e4edefa223490fcdd6c4ff272a 2025-07-17T08:34:52.1464391Z * [new tag] trunk/898179331e82e5ee1ff64737f508fd6a6da711f9 -> trunk/898179331e82e5ee1ff64737f508fd6a6da711f9 2025-07-17T08:34:52.1466925Z * [new tag] trunk/899d3d3e9eec4756a2e759f0203334dbe08f664e -> trunk/899d3d3e9eec4756a2e759f0203334dbe08f664e 2025-07-17T08:34:52.1469220Z * [new tag] trunk/8a2255130025abc80e03adaeb0b90d6463f2f916 -> trunk/8a2255130025abc80e03adaeb0b90d6463f2f916 2025-07-17T08:34:52.1471288Z * [new tag] trunk/8a396c56350ade97c18a6e13b3d97939f1eea075 -> trunk/8a396c56350ade97c18a6e13b3d97939f1eea075 2025-07-17T08:34:52.1473606Z * [new tag] trunk/8a47f9d03be0e0118a43443ff5f60db7be85102b -> trunk/8a47f9d03be0e0118a43443ff5f60db7be85102b 2025-07-17T08:34:52.1477361Z * [new tag] trunk/8a88c6e85abe345d95d970d424a0295cbc89b69c -> trunk/8a88c6e85abe345d95d970d424a0295cbc89b69c 2025-07-17T08:34:52.1479587Z * [new tag] trunk/8a8fac11318778aa5e4479b369b37fdb9e10ec17 -> trunk/8a8fac11318778aa5e4479b369b37fdb9e10ec17 2025-07-17T08:34:52.1481951Z * [new tag] trunk/8ad6197b466e91eb5fbbde2c76688d40fa472ac3 -> trunk/8ad6197b466e91eb5fbbde2c76688d40fa472ac3 2025-07-17T08:34:52.1484219Z * [new tag] trunk/8b0e0e4f2301f4aeb9fccf52cd2c6d8ac13da131 -> trunk/8b0e0e4f2301f4aeb9fccf52cd2c6d8ac13da131 2025-07-17T08:34:52.1486459Z * [new tag] trunk/8b68e5b1bb4a0014a64eab25c54b716968e109a9 -> trunk/8b68e5b1bb4a0014a64eab25c54b716968e109a9 2025-07-17T08:34:52.1488722Z * [new tag] trunk/8b8684466a11ae6ac27c85dbc6a02091dadac749 -> trunk/8b8684466a11ae6ac27c85dbc6a02091dadac749 2025-07-17T08:34:52.1491031Z * [new tag] trunk/8b97e4dd8cbe23101801043fe343d7350e527540 -> trunk/8b97e4dd8cbe23101801043fe343d7350e527540 2025-07-17T08:34:52.1493545Z * [new tag] trunk/8bda95228fbefa6ce204bf4da8b632d1516431bb -> trunk/8bda95228fbefa6ce204bf4da8b632d1516431bb 2025-07-17T08:34:52.1495754Z * [new tag] trunk/8c0df6fe1769c813f57da766528ff3cc4e56943e -> trunk/8c0df6fe1769c813f57da766528ff3cc4e56943e 2025-07-17T08:34:52.1498102Z * [new tag] trunk/8c2e45008282cf5202b72a0ecb0c2951438abeea -> trunk/8c2e45008282cf5202b72a0ecb0c2951438abeea 2025-07-17T08:34:52.1500408Z * [new tag] trunk/8c3f206457a1b9d75bc95a6c30a101135fcee329 -> trunk/8c3f206457a1b9d75bc95a6c30a101135fcee329 2025-07-17T08:34:52.1502850Z * [new tag] trunk/8c5b070d1f7cf4a9b58b09a28a0aeb5fdca28679 -> trunk/8c5b070d1f7cf4a9b58b09a28a0aeb5fdca28679 2025-07-17T08:34:52.1505045Z * [new tag] trunk/8c928372b3560bb512e2cf98c94efd923caf5fa7 -> trunk/8c928372b3560bb512e2cf98c94efd923caf5fa7 2025-07-17T08:34:52.1507381Z * [new tag] trunk/8cb0c4a4da0e017834d63f3d7dc2ab29e0ad6424 -> trunk/8cb0c4a4da0e017834d63f3d7dc2ab29e0ad6424 2025-07-17T08:34:52.1509620Z * [new tag] trunk/8d070187e34d5474ae88cefe56754f729138756b -> trunk/8d070187e34d5474ae88cefe56754f729138756b 2025-07-17T08:34:52.1511821Z * [new tag] trunk/8d7ee0f4fb7e41e8a49a5a24b6686b2324c493ba -> trunk/8d7ee0f4fb7e41e8a49a5a24b6686b2324c493ba 2025-07-17T08:34:52.1514224Z * [new tag] trunk/8da774d81feace041ff39de4ba2ddd43086370d3 -> trunk/8da774d81feace041ff39de4ba2ddd43086370d3 2025-07-17T08:34:52.1516567Z * [new tag] trunk/8dff457f42c9a20f9936d22773239df33cb48c9d -> trunk/8dff457f42c9a20f9936d22773239df33cb48c9d 2025-07-17T08:34:52.1518812Z * [new tag] trunk/8e02cd9c5ab151faad80e18e648c82a204fc3735 -> trunk/8e02cd9c5ab151faad80e18e648c82a204fc3735 2025-07-17T08:34:52.1521063Z * [new tag] trunk/8e8bbfc80340f0bf0a31a03d61b9b4dd72a40469 -> trunk/8e8bbfc80340f0bf0a31a03d61b9b4dd72a40469 2025-07-17T08:34:52.1523171Z * [new tag] trunk/8e8d4b13b0f6358e249a1e6fea9831b91b41b6d0 -> trunk/8e8d4b13b0f6358e249a1e6fea9831b91b41b6d0 2025-07-17T08:34:52.1525549Z * [new tag] trunk/8eaa9f2701277f328d9d6aea1bfe7cba20792f7c -> trunk/8eaa9f2701277f328d9d6aea1bfe7cba20792f7c 2025-07-17T08:34:52.1527840Z * [new tag] trunk/8eb3c5b7a18205974fff0353b97e599ea3d0e2a4 -> trunk/8eb3c5b7a18205974fff0353b97e599ea3d0e2a4 2025-07-17T08:34:52.1530231Z * [new tag] trunk/8f02161d1012143263fdbca47ee62983448e2c7e -> trunk/8f02161d1012143263fdbca47ee62983448e2c7e 2025-07-17T08:34:52.1532698Z * [new tag] trunk/8f0998aafe5e9594ade50cd263d5c63a5aad9333 -> trunk/8f0998aafe5e9594ade50cd263d5c63a5aad9333 2025-07-17T08:34:52.1534986Z * [new tag] trunk/8f5f01bf19e01f4b74add7f3d558966152a398c7 -> trunk/8f5f01bf19e01f4b74add7f3d558966152a398c7 2025-07-17T08:34:52.1537225Z * [new tag] trunk/8f9a191db6efcf44510db87c8f6f6e96c67e9b11 -> trunk/8f9a191db6efcf44510db87c8f6f6e96c67e9b11 2025-07-17T08:34:52.1539496Z * [new tag] trunk/8fcda2c60d974e4970ce939bc828bece9903c304 -> trunk/8fcda2c60d974e4970ce939bc828bece9903c304 2025-07-17T08:34:52.1541606Z * [new tag] trunk/900fba4c073b121b6c9ce581ea27e25c13a354e5 -> trunk/900fba4c073b121b6c9ce581ea27e25c13a354e5 2025-07-17T08:34:52.1543883Z * [new tag] trunk/9056279f8159b052599a31b591a78da1acc4224c -> trunk/9056279f8159b052599a31b591a78da1acc4224c 2025-07-17T08:34:52.1546155Z * [new tag] trunk/905b0846903a0193ee651a44762e72c881f83950 -> trunk/905b0846903a0193ee651a44762e72c881f83950 2025-07-17T08:34:52.1548522Z * [new tag] trunk/905b194a2e3d1d1c9ec136dfc1b6be7e62ff4b2c -> trunk/905b194a2e3d1d1c9ec136dfc1b6be7e62ff4b2c 2025-07-17T08:34:52.1550930Z * [new tag] trunk/90618581e971d28ac6950305d72521af05ed3a42 -> trunk/90618581e971d28ac6950305d72521af05ed3a42 2025-07-17T08:34:52.1553172Z * [new tag] trunk/907aea032d2f709f26dec78b762481c4604cc186 -> trunk/907aea032d2f709f26dec78b762481c4604cc186 2025-07-17T08:34:52.1555683Z * [new tag] trunk/907d0931cc8868164c2890b0452eda5d4da49278 -> trunk/907d0931cc8868164c2890b0452eda5d4da49278 2025-07-17T08:34:52.1558061Z * [new tag] trunk/90b973a2e228f2a5e2c158e5f5cf8f59d9118fc8 -> trunk/90b973a2e228f2a5e2c158e5f5cf8f59d9118fc8 2025-07-17T08:34:52.1560370Z * [new tag] trunk/91ee9ee82d40bcc36afbe9fed8dfc2f8c19c7316 -> trunk/91ee9ee82d40bcc36afbe9fed8dfc2f8c19c7316 2025-07-17T08:34:52.1562581Z * [new tag] trunk/9222552572890fd7ff1d75884cc6a99a55ab90bc -> trunk/9222552572890fd7ff1d75884cc6a99a55ab90bc 2025-07-17T08:34:52.1564827Z * [new tag] trunk/92388bb2ab14dae7b63c2088de4d6add413d1c27 -> trunk/92388bb2ab14dae7b63c2088de4d6add413d1c27 2025-07-17T08:34:52.1567247Z * [new tag] trunk/92409b6c89fbfbd3caa79c81b1e3d9e7917d3bc7 -> trunk/92409b6c89fbfbd3caa79c81b1e3d9e7917d3bc7 2025-07-17T08:34:52.1569683Z * [new tag] trunk/924fc52e1842de756a7632e7b2777aeed77d60e5 -> trunk/924fc52e1842de756a7632e7b2777aeed77d60e5 2025-07-17T08:34:52.1571917Z * [new tag] trunk/925fbfca27580f2ec8d5f9bfd549b1fe2e665617 -> trunk/925fbfca27580f2ec8d5f9bfd549b1fe2e665617 2025-07-17T08:34:52.1574148Z * [new tag] trunk/92b7ed6d07fcd6af31cddc7b4276e626b023bd0b -> trunk/92b7ed6d07fcd6af31cddc7b4276e626b023bd0b 2025-07-17T08:34:52.1576566Z * [new tag] trunk/92c79f36dbbdb01e0aab6efb8f52df828e1046d5 -> trunk/92c79f36dbbdb01e0aab6efb8f52df828e1046d5 2025-07-17T08:34:52.1578702Z * [new tag] trunk/92ee5bd9f6bef38b3e3e81a37f4d5374395aeb40 -> trunk/92ee5bd9f6bef38b3e3e81a37f4d5374395aeb40 2025-07-17T08:34:52.1581085Z * [new tag] trunk/92f41ccc2651ab284118ec7087977ab455027eb0 -> trunk/92f41ccc2651ab284118ec7087977ab455027eb0 2025-07-17T08:34:52.1583336Z * [new tag] trunk/930b575389f9233efddf70ea7b7804ed06af80d5 -> trunk/930b575389f9233efddf70ea7b7804ed06af80d5 2025-07-17T08:34:52.1585775Z * [new tag] trunk/9328a7fb589f6372ea478e47cdae8d124030f383 -> trunk/9328a7fb589f6372ea478e47cdae8d124030f383 2025-07-17T08:34:52.1588148Z * [new tag] trunk/9338d85d4594f981c198bcbb7edfccf6b92643be -> trunk/9338d85d4594f981c198bcbb7edfccf6b92643be 2025-07-17T08:34:52.1590290Z * [new tag] trunk/9345279c6ebdbad95b7b53bc2cb6f63a4e57b2cc -> trunk/9345279c6ebdbad95b7b53bc2cb6f63a4e57b2cc 2025-07-17T08:34:52.1592372Z * [new tag] trunk/937529f0b31788726e53890f5601886c64dc9eec -> trunk/937529f0b31788726e53890f5601886c64dc9eec 2025-07-17T08:34:52.1594800Z * [new tag] trunk/938515fa75f4e9d28c62a2aaa3e0e5428c867c36 -> trunk/938515fa75f4e9d28c62a2aaa3e0e5428c867c36 2025-07-17T08:34:52.1597172Z * [new tag] trunk/93854e83b7bfde94090662e9b372d8bf44ccf5d4 -> trunk/93854e83b7bfde94090662e9b372d8bf44ccf5d4 2025-07-17T08:34:52.1599401Z * [new tag] trunk/944a140e90389eced1ec38e14cb4345811ed0b1a -> trunk/944a140e90389eced1ec38e14cb4345811ed0b1a 2025-07-17T08:34:52.1601660Z * [new tag] trunk/9462106b7e41d9a24308255e15613b464ab086ce -> trunk/9462106b7e41d9a24308255e15613b464ab086ce 2025-07-17T08:34:52.1603893Z * [new tag] trunk/94716db22214912896cf680dc3eb88574f611a42 -> trunk/94716db22214912896cf680dc3eb88574f611a42 2025-07-17T08:34:52.1606240Z * [new tag] trunk/94763f5ca75330206777f3c1b36ff124706bd9d9 -> trunk/94763f5ca75330206777f3c1b36ff124706bd9d9 2025-07-17T08:34:52.1608540Z * [new tag] trunk/94995eba07763890b86465d53c4647c089f48d0a -> trunk/94995eba07763890b86465d53c4647c089f48d0a 2025-07-17T08:34:52.1610802Z * [new tag] trunk/94ae61533719defecb08d11f63c5d9fe9a826c00 -> trunk/94ae61533719defecb08d11f63c5d9fe9a826c00 2025-07-17T08:34:52.1613115Z * [new tag] trunk/94c746bb43484787a3f5bbdc2f72bd4fb02f2964 -> trunk/94c746bb43484787a3f5bbdc2f72bd4fb02f2964 2025-07-17T08:34:52.1615558Z * [new tag] trunk/94da4523ec4362dabbf9f08f9f2efce79ab70b73 -> trunk/94da4523ec4362dabbf9f08f9f2efce79ab70b73 2025-07-17T08:34:52.1617623Z * [new tag] trunk/94f8679019ea4b1272f1ad58ad7cad87147cf5a7 -> trunk/94f8679019ea4b1272f1ad58ad7cad87147cf5a7 2025-07-17T08:34:52.1619829Z * [new tag] trunk/9508d73307b4bc1fe453677526a096e5e10a7575 -> trunk/9508d73307b4bc1fe453677526a096e5e10a7575 2025-07-17T08:34:52.1622106Z * [new tag] trunk/9513b9d03fa8950ba5d2b59cc0b1a1aab3a41c06 -> trunk/9513b9d03fa8950ba5d2b59cc0b1a1aab3a41c06 2025-07-17T08:34:52.1624441Z * [new tag] trunk/95448b2ce61c3995ecdae0bfec655a5ef81a117d -> trunk/95448b2ce61c3995ecdae0bfec655a5ef81a117d 2025-07-17T08:34:52.1626821Z * [new tag] trunk/954ce949500746763a487a8ed9800035e7cd87d4 -> trunk/954ce949500746763a487a8ed9800035e7cd87d4 2025-07-17T08:34:52.1629089Z * [new tag] trunk/95a7d1912a86d857a9f4adb55b73b57518ce1e7b -> trunk/95a7d1912a86d857a9f4adb55b73b57518ce1e7b 2025-07-17T08:34:52.1631350Z * [new tag] trunk/95bc3da9f8555897d5d2b90f75385e3b2ff7bc43 -> trunk/95bc3da9f8555897d5d2b90f75385e3b2ff7bc43 2025-07-17T08:34:52.1633566Z * [new tag] trunk/95cb42c45d17f532222611e8028c9307622cc3c9 -> trunk/95cb42c45d17f532222611e8028c9307622cc3c9 2025-07-17T08:34:52.1635957Z * [new tag] trunk/9620994067b18e846a097d1e99af85ec2426ef0a -> trunk/9620994067b18e846a097d1e99af85ec2426ef0a 2025-07-17T08:34:52.1638236Z * [new tag] trunk/9636e2cfd3e995ef977f670ad47e8e895296d992 -> trunk/9636e2cfd3e995ef977f670ad47e8e895296d992 2025-07-17T08:34:52.1640466Z * [new tag] trunk/9642c7568967ab424c5d0e04ef2cd1e82a54b5f8 -> trunk/9642c7568967ab424c5d0e04ef2cd1e82a54b5f8 2025-07-17T08:34:52.1642855Z * [new tag] trunk/9656251bb1bee12c0e2f21828dc14a4c3c06afdd -> trunk/9656251bb1bee12c0e2f21828dc14a4c3c06afdd 2025-07-17T08:34:52.1645152Z * [new tag] trunk/9665702c64af633ab23499228d552a49660a9fa2 -> trunk/9665702c64af633ab23499228d552a49660a9fa2 2025-07-17T08:34:52.1647426Z * [new tag] trunk/96897e721b76f3b72aa406696165711d74f8d260 -> trunk/96897e721b76f3b72aa406696165711d74f8d260 2025-07-17T08:34:52.1649778Z * [new tag] trunk/968f90ce7344223c788eeba59200cc22b9f94dcd -> trunk/968f90ce7344223c788eeba59200cc22b9f94dcd 2025-07-17T08:34:52.1652033Z * [new tag] trunk/96d082d06bda98addd4ad7903d315477404dc272 -> trunk/96d082d06bda98addd4ad7903d315477404dc272 2025-07-17T08:34:52.1654193Z * [new tag] trunk/96df86641048a282a4622ac93d19e4c7b6c38b8e -> trunk/96df86641048a282a4622ac93d19e4c7b6c38b8e 2025-07-17T08:34:52.1656539Z * [new tag] trunk/96e4c95cd8d03037765ffd4b8fcdaa06c4b2c51c -> trunk/96e4c95cd8d03037765ffd4b8fcdaa06c4b2c51c 2025-07-17T08:34:52.1658796Z * [new tag] trunk/9768d393fa62df8a508136f5b6634bf955d8365d -> trunk/9768d393fa62df8a508136f5b6634bf955d8365d 2025-07-17T08:34:52.1661175Z * [new tag] trunk/977abe786d907c1ff76528a550e3d53c9f3b1044 -> trunk/977abe786d907c1ff76528a550e3d53c9f3b1044 2025-07-17T08:34:52.1663460Z * [new tag] trunk/981c99fdffd9a4f510e4b89245d16aa427ee0978 -> trunk/981c99fdffd9a4f510e4b89245d16aa427ee0978 2025-07-17T08:34:52.1665736Z * [new tag] trunk/987314aa96fdf8aa051e3643b26f4209b7fe166d -> trunk/987314aa96fdf8aa051e3643b26f4209b7fe166d 2025-07-17T08:34:52.1668000Z * [new tag] trunk/98a34e8d4b4d73504afbf49f70284221e6303314 -> trunk/98a34e8d4b4d73504afbf49f70284221e6303314 2025-07-17T08:34:52.1670191Z * [new tag] trunk/98bb0c0e7888a84de1d22c4952de69a5d5b1a0b3 -> trunk/98bb0c0e7888a84de1d22c4952de69a5d5b1a0b3 2025-07-17T08:34:52.1672533Z * [new tag] trunk/9944cd0949e96d656746de6df0cb3a5a954597e4 -> trunk/9944cd0949e96d656746de6df0cb3a5a954597e4 2025-07-17T08:34:52.1674995Z * [new tag] trunk/996206e66fcafff25d0af5177497e8f792000869 -> trunk/996206e66fcafff25d0af5177497e8f792000869 2025-07-17T08:34:52.1677474Z * [new tag] trunk/9963845a4e55a3e8d35ea49deb8f8c1f6c53cb74 -> trunk/9963845a4e55a3e8d35ea49deb8f8c1f6c53cb74 2025-07-17T08:34:52.1679711Z * [new tag] trunk/9968c854b6a38857d05fb1da7bea797fbf23c880 -> trunk/9968c854b6a38857d05fb1da7bea797fbf23c880 2025-07-17T08:34:52.1681930Z * [new tag] trunk/9968edd00256fdb47e2e0129df918c5b23c06419 -> trunk/9968edd00256fdb47e2e0129df918c5b23c06419 2025-07-17T08:34:52.1684214Z * [new tag] trunk/99c1a6bdd98ecb1f9b2fb8310bcd891c8c6adf57 -> trunk/99c1a6bdd98ecb1f9b2fb8310bcd891c8c6adf57 2025-07-17T08:34:52.1686515Z * [new tag] trunk/99e99d5bfe57eddfce23f7dc9ede016cb1393c52 -> trunk/99e99d5bfe57eddfce23f7dc9ede016cb1393c52 2025-07-17T08:34:52.1688892Z * [new tag] trunk/9a2c669425379eb264f896390b8fcd8d3f2ce959 -> trunk/9a2c669425379eb264f896390b8fcd8d3f2ce959 2025-07-17T08:34:52.1691128Z * [new tag] trunk/9a42f01586267b52fe303c0c1b7a71d50cc4178e -> trunk/9a42f01586267b52fe303c0c1b7a71d50cc4178e 2025-07-17T08:34:52.1693372Z * [new tag] trunk/9a4c08ddfc9b43c07cd16355277d359dfcef50d6 -> trunk/9a4c08ddfc9b43c07cd16355277d359dfcef50d6 2025-07-17T08:34:52.1695653Z * [new tag] trunk/9a5278225fc5e7b46d54a65ae1a3f049ee49824f -> trunk/9a5278225fc5e7b46d54a65ae1a3f049ee49824f 2025-07-17T08:34:52.1698521Z * [new tag] trunk/9a5c59368d46d3793e09b72a2782ef4d12e74392 -> trunk/9a5c59368d46d3793e09b72a2782ef4d12e74392 2025-07-17T08:34:52.1700784Z * [new tag] trunk/9aaa184105b2f436b5834187c4c004c02e438491 -> trunk/9aaa184105b2f436b5834187c4c004c02e438491 2025-07-17T08:34:52.1703077Z * [new tag] trunk/9afee0fa969fd4ebab86a2cbad3217c8bade156e -> trunk/9afee0fa969fd4ebab86a2cbad3217c8bade156e 2025-07-17T08:34:52.1705331Z * [new tag] trunk/9b0013c6bb98d7161e921d03be76c81bbc0eebef -> trunk/9b0013c6bb98d7161e921d03be76c81bbc0eebef 2025-07-17T08:34:52.1707587Z * [new tag] trunk/9b122aab5dcd419b96d4a1137c45e99bde94f14e -> trunk/9b122aab5dcd419b96d4a1137c45e99bde94f14e 2025-07-17T08:34:52.1709906Z * [new tag] trunk/9b498d3bb28b8e3411ce464dd2755c5b96d92c8f -> trunk/9b498d3bb28b8e3411ce464dd2755c5b96d92c8f 2025-07-17T08:34:52.1712120Z * [new tag] trunk/9b4a748e29a720d0fade7e1298a68cc36cfd5f5e -> trunk/9b4a748e29a720d0fade7e1298a68cc36cfd5f5e 2025-07-17T08:34:52.1714561Z * [new tag] trunk/9b4db093cb63555654c868ba9e3ab68ef4c722e5 -> trunk/9b4db093cb63555654c868ba9e3ab68ef4c722e5 2025-07-17T08:34:52.1717049Z * [new tag] trunk/9bae2fcf990cd51891218895ac1b3cfe70ffc2c3 -> trunk/9bae2fcf990cd51891218895ac1b3cfe70ffc2c3 2025-07-17T08:34:52.1719368Z * [new tag] trunk/9bd0830ed8d3e753afc3e9fff2553f4c5871f6f9 -> trunk/9bd0830ed8d3e753afc3e9fff2553f4c5871f6f9 2025-07-17T08:34:52.1721619Z * [new tag] trunk/9bd42c15707a4b410ee005d5916e882a7db432bb -> trunk/9bd42c15707a4b410ee005d5916e882a7db432bb 2025-07-17T08:34:52.1724274Z * [new tag] trunk/9bdf87e8918b9a3f78d7bcb8a770c19f7c82ac15 -> trunk/9bdf87e8918b9a3f78d7bcb8a770c19f7c82ac15 2025-07-17T08:34:52.1726723Z * [new tag] trunk/9be5860bc34b643abd02ac5b5a89d820efcdfbc9 -> trunk/9be5860bc34b643abd02ac5b5a89d820efcdfbc9 2025-07-17T08:34:52.1729294Z * [new tag] trunk/9bf41633d76f7816fdcc0b7ae22a42dc91821004 -> trunk/9bf41633d76f7816fdcc0b7ae22a42dc91821004 2025-07-17T08:34:52.1731544Z * [new tag] trunk/9bf6593e96b711641606e6008a4936173fd3b458 -> trunk/9bf6593e96b711641606e6008a4936173fd3b458 2025-07-17T08:34:52.1733895Z * [new tag] trunk/9bfefda296d2df31ce9bb90410d91d8f18127545 -> trunk/9bfefda296d2df31ce9bb90410d91d8f18127545 2025-07-17T08:34:52.1736344Z * [new tag] trunk/9c189ed29a2cfa08ff357cb5042ff8c34f07d668 -> trunk/9c189ed29a2cfa08ff357cb5042ff8c34f07d668 2025-07-17T08:34:52.1738851Z * [new tag] trunk/9c39bc24807a5843f8affdf56bd71836760dc554 -> trunk/9c39bc24807a5843f8affdf56bd71836760dc554 2025-07-17T08:34:52.1740951Z * [new tag] trunk/9ca080db87f66becb2bb2c61598c0beabae73d5f -> trunk/9ca080db87f66becb2bb2c61598c0beabae73d5f 2025-07-17T08:34:52.1743265Z * [new tag] trunk/9cced33c7c9cdfc68e9ed85c4b45bdfd05dac92d -> trunk/9cced33c7c9cdfc68e9ed85c4b45bdfd05dac92d 2025-07-17T08:34:52.1745519Z * [new tag] trunk/9cd521de4dad5fc6bca94e253a9334b9a521acb0 -> trunk/9cd521de4dad5fc6bca94e253a9334b9a521acb0 2025-07-17T08:34:52.1747747Z * [new tag] trunk/9d175bc7e615b062d8ea57df6ca17edb8a6b951f -> trunk/9d175bc7e615b062d8ea57df6ca17edb8a6b951f 2025-07-17T08:34:52.1750008Z * [new tag] trunk/9d184bda2f190a3ba72a4a0d95e1fde26d6bfc08 -> trunk/9d184bda2f190a3ba72a4a0d95e1fde26d6bfc08 2025-07-17T08:34:52.1752322Z * [new tag] trunk/9d2d2270037c5a014767cdfa531863da8062bf9d -> trunk/9d2d2270037c5a014767cdfa531863da8062bf9d 2025-07-17T08:34:52.1755049Z * [new tag] trunk/9d59b516e9b3026948918e3ff8c2ef55a33d13ad -> trunk/9d59b516e9b3026948918e3ff8c2ef55a33d13ad 2025-07-17T08:34:52.1757333Z * [new tag] trunk/9d677389cb5eb75d423860c55519b522961a9195 -> trunk/9d677389cb5eb75d423860c55519b522961a9195 2025-07-17T08:34:52.1759615Z * [new tag] trunk/9de23d0c29dfac8dc0f6f234bdbcd85a6375fa81 -> trunk/9de23d0c29dfac8dc0f6f234bdbcd85a6375fa81 2025-07-17T08:34:52.1761841Z * [new tag] trunk/9df0176408518b30ac172837bd697c9d19b19a98 -> trunk/9df0176408518b30ac172837bd697c9d19b19a98 2025-07-17T08:34:52.1764144Z * [new tag] trunk/9df2e8020f190bdad02258aff2b2f88f7db3dcab -> trunk/9df2e8020f190bdad02258aff2b2f88f7db3dcab 2025-07-17T08:34:52.1766443Z * [new tag] trunk/9e132b770e06c2399f827d3d1963fe036e9744e6 -> trunk/9e132b770e06c2399f827d3d1963fe036e9744e6 2025-07-17T08:34:52.1768562Z * [new tag] trunk/9e5f4a844c0aebf964a435094005c92713fbe99a -> trunk/9e5f4a844c0aebf964a435094005c92713fbe99a 2025-07-17T08:34:52.1770840Z * [new tag] trunk/9e88d6c857fd2b276427f4527130e2ecb7ed731e -> trunk/9e88d6c857fd2b276427f4527130e2ecb7ed731e 2025-07-17T08:34:52.1773277Z * [new tag] trunk/9e9484d022f8d65e2066b5683b75c84ba0f36332 -> trunk/9e9484d022f8d65e2066b5683b75c84ba0f36332 2025-07-17T08:34:52.1775739Z * [new tag] trunk/9ed0060225a7b78c60c42f29be94444b537edd4a -> trunk/9ed0060225a7b78c60c42f29be94444b537edd4a 2025-07-17T08:34:52.1777970Z * [new tag] trunk/9f18482d41227df3cf2248dfa54bd6601e61e1ca -> trunk/9f18482d41227df3cf2248dfa54bd6601e61e1ca 2025-07-17T08:34:52.1780402Z * [new tag] trunk/9f37cce69334bccebf4b21503f0047d0c0bb320c -> trunk/9f37cce69334bccebf4b21503f0047d0c0bb320c 2025-07-17T08:34:52.1782600Z * [new tag] trunk/9f5153b1a495b3b2b7ad9f22a064fe4074d742aa -> trunk/9f5153b1a495b3b2b7ad9f22a064fe4074d742aa 2025-07-17T08:34:52.1784846Z * [new tag] trunk/9f5276dc07c788533af8945b1605df47a33313e7 -> trunk/9f5276dc07c788533af8945b1605df47a33313e7 2025-07-17T08:34:52.1787163Z * [new tag] trunk/9fe2d156a9f81d67e248c0edaf7feee1a8d6c4d5 -> trunk/9fe2d156a9f81d67e248c0edaf7feee1a8d6c4d5 2025-07-17T08:34:52.1789485Z * [new tag] trunk/9fed2addedb42da86b657165fe14eadc911232cf -> trunk/9fed2addedb42da86b657165fe14eadc911232cf 2025-07-17T08:34:52.1791692Z * [new tag] trunk/a00a697c1748937b0c6d13508ff76b8af4957d8a -> trunk/a00a697c1748937b0c6d13508ff76b8af4957d8a 2025-07-17T08:34:52.1793980Z * [new tag] trunk/a0308edb6cdfd8983e80a499890d9f320556e844 -> trunk/a0308edb6cdfd8983e80a499890d9f320556e844 2025-07-17T08:34:52.1796353Z * [new tag] trunk/a04a13c44908fe0ace4f76a228d045dbf5c015bc -> trunk/a04a13c44908fe0ace4f76a228d045dbf5c015bc 2025-07-17T08:34:52.1798830Z * [new tag] trunk/a0e0abd0379ed403adc926d3cc5459fa35aff0a0 -> trunk/a0e0abd0379ed403adc926d3cc5459fa35aff0a0 2025-07-17T08:34:52.1801289Z * [new tag] trunk/a10024d7dea47c52469059a47efe376eb20adca0 -> trunk/a10024d7dea47c52469059a47efe376eb20adca0 2025-07-17T08:34:52.1803489Z * [new tag] trunk/a1057cda31fe890d237fe9a2e6b3314cd4be3439 -> trunk/a1057cda31fe890d237fe9a2e6b3314cd4be3439 2025-07-17T08:34:52.1805932Z * [new tag] trunk/a1257446f8eaf81eb758e89b4e784bed47b9c733 -> trunk/a1257446f8eaf81eb758e89b4e784bed47b9c733 2025-07-17T08:34:52.1808415Z * [new tag] trunk/a1282b18239204b0344884ebd232b33b2d8b748f -> trunk/a1282b18239204b0344884ebd232b33b2d8b748f 2025-07-17T08:34:52.1810563Z * [new tag] trunk/a14f427db68e54500ef4cd9ed34cb9537263bb74 -> trunk/a14f427db68e54500ef4cd9ed34cb9537263bb74 2025-07-17T08:34:52.1812844Z * [new tag] trunk/a1dad2f2d2c082e2a3784c3d585ef0204b7ccf75 -> trunk/a1dad2f2d2c082e2a3784c3d585ef0204b7ccf75 2025-07-17T08:34:52.1815088Z * [new tag] trunk/a1e4f1f98a0b9596fe52aaf2f85b0778498d5f49 -> trunk/a1e4f1f98a0b9596fe52aaf2f85b0778498d5f49 2025-07-17T08:34:52.1817381Z * [new tag] trunk/a205e8fd7357cf5c2b95eb4bec7c8ff7b5c41651 -> trunk/a205e8fd7357cf5c2b95eb4bec7c8ff7b5c41651 2025-07-17T08:34:52.1819659Z * [new tag] trunk/a21806f0386c2671fc11ae37fe2fe96c4aeb37d9 -> trunk/a21806f0386c2671fc11ae37fe2fe96c4aeb37d9 2025-07-17T08:34:52.1821869Z * [new tag] trunk/a23f4471b952d8cd630b860639e0aaa9be957d60 -> trunk/a23f4471b952d8cd630b860639e0aaa9be957d60 2025-07-17T08:34:52.1824022Z * [new tag] trunk/a24afbff3f44399ac9513addce6bd50cacab22d2 -> trunk/a24afbff3f44399ac9513addce6bd50cacab22d2 2025-07-17T08:34:52.1826109Z * [new tag] trunk/a24ce67deedab025531660ddb44c148bb066edf8 -> trunk/a24ce67deedab025531660ddb44c148bb066edf8 2025-07-17T08:34:52.1828444Z * [new tag] trunk/a25d1443fae79a9984eaa44e973b2f5645c85da8 -> trunk/a25d1443fae79a9984eaa44e973b2f5645c85da8 2025-07-17T08:34:52.1830745Z * [new tag] trunk/a26bf3892778ca7cc457c772a1f5194c11b6f33c -> trunk/a26bf3892778ca7cc457c772a1f5194c11b6f33c 2025-07-17T08:34:52.1833034Z * [new tag] trunk/a28e6ae38f1e1de33ecf48b679ff1eda71fc2efc -> trunk/a28e6ae38f1e1de33ecf48b679ff1eda71fc2efc 2025-07-17T08:34:52.1835388Z * [new tag] trunk/a2a75be0f8d680cf1dc36c50f34f225b0c7c4d6a -> trunk/a2a75be0f8d680cf1dc36c50f34f225b0c7c4d6a 2025-07-17T08:34:52.1837718Z * [new tag] trunk/a2ad16be72bf989c01b96c4c56b1c108a71c087f -> trunk/a2ad16be72bf989c01b96c4c56b1c108a71c087f 2025-07-17T08:34:52.1840020Z * [new tag] trunk/a2b0b2698d5b953861b2e4f3cdee11136f07bd3b -> trunk/a2b0b2698d5b953861b2e4f3cdee11136f07bd3b 2025-07-17T08:34:52.1842283Z * [new tag] trunk/a3098a74d494020dbb906c05ef047013e1921662 -> trunk/a3098a74d494020dbb906c05ef047013e1921662 2025-07-17T08:34:52.1844828Z * [new tag] trunk/a317c63d1b70a7989a4cab09394691fb7a3d8323 -> trunk/a317c63d1b70a7989a4cab09394691fb7a3d8323 2025-07-17T08:34:52.1846796Z * [new tag] trunk/a355158fcba807fda1e47e5ee42babdbcf447947 -> trunk/a355158fcba807fda1e47e5ee42babdbcf447947 2025-07-17T08:34:52.1849058Z * [new tag] trunk/a35b3a9b95bbe0703de33802162a83cb7970ddc0 -> trunk/a35b3a9b95bbe0703de33802162a83cb7970ddc0 2025-07-17T08:34:52.1851478Z * [new tag] trunk/a369350065493109d1abfbb994695777ab11bcf4 -> trunk/a369350065493109d1abfbb994695777ab11bcf4 2025-07-17T08:34:52.1853673Z * [new tag] trunk/a38f433be2e94a64b095a44ba39879d02d0c2316 -> trunk/a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:34:52.1855888Z * [new tag] trunk/a3ec6d64b24fdbd61164240f3bcd8530434812b5 -> trunk/a3ec6d64b24fdbd61164240f3bcd8530434812b5 2025-07-17T08:34:52.1858184Z * [new tag] trunk/a46ea8a364e528828c4369d58d53802291f0b49b -> trunk/a46ea8a364e528828c4369d58d53802291f0b49b 2025-07-17T08:34:52.1860728Z * [new tag] trunk/a47ca4fc746a663c0e97d55a87815d0965d0a7e9 -> trunk/a47ca4fc746a663c0e97d55a87815d0965d0a7e9 2025-07-17T08:34:52.1862758Z * [new tag] trunk/a4b59498c5222a02a706fa1e1308a6035d65feab -> trunk/a4b59498c5222a02a706fa1e1308a6035d65feab 2025-07-17T08:34:52.1865150Z * [new tag] trunk/a4c7e7f98373ad8f309e419c6f98b0134933dcda -> trunk/a4c7e7f98373ad8f309e419c6f98b0134933dcda 2025-07-17T08:34:52.1867473Z * [new tag] trunk/a4d753295ee5662056bdfd1b00fa242071ac7125 -> trunk/a4d753295ee5662056bdfd1b00fa242071ac7125 2025-07-17T08:34:52.1869958Z * [new tag] trunk/a4ea242edc8626250e54a62aa073cba0e5322af2 -> trunk/a4ea242edc8626250e54a62aa073cba0e5322af2 2025-07-17T08:34:52.1872430Z * [new tag] trunk/a529a5daf5d348f4f8dac4fcd85521153b6a3d4d -> trunk/a529a5daf5d348f4f8dac4fcd85521153b6a3d4d 2025-07-17T08:34:52.1874789Z * [new tag] trunk/a5938ff431202864c92fec7d9042574348b1c15b -> trunk/a5938ff431202864c92fec7d9042574348b1c15b 2025-07-17T08:34:52.1877188Z * [new tag] trunk/a5b4463d60e5beaec546843b24876ce573890784 -> trunk/a5b4463d60e5beaec546843b24876ce573890784 2025-07-17T08:34:52.1879317Z * [new tag] trunk/a5c61eb78dda6b7b4c4beb8bc616f9545ea4807b -> trunk/a5c61eb78dda6b7b4c4beb8bc616f9545ea4807b 2025-07-17T08:34:52.1881676Z * [new tag] trunk/a5cbb2bcb3c005909a3c10042605e861e6390927 -> trunk/a5cbb2bcb3c005909a3c10042605e861e6390927 2025-07-17T08:34:52.1884039Z * [new tag] trunk/a5df6ffbc2cf8d61fc98bca1c10f22ff64b2318e -> trunk/a5df6ffbc2cf8d61fc98bca1c10f22ff64b2318e 2025-07-17T08:34:52.1886347Z * [new tag] trunk/a5e68814d556cf67c6511876410970dd08c3dd6d -> trunk/a5e68814d556cf67c6511876410970dd08c3dd6d 2025-07-17T08:34:52.1888638Z * [new tag] trunk/a5f59cc2eab3a5201712c52fe48c268357ba4f3c -> trunk/a5f59cc2eab3a5201712c52fe48c268357ba4f3c 2025-07-17T08:34:52.1891034Z * [new tag] trunk/a6084b71edb8d2856356724b5e71c4e2a861867f -> trunk/a6084b71edb8d2856356724b5e71c4e2a861867f 2025-07-17T08:34:52.1893316Z * [new tag] trunk/a6210fd07b8fe1924f24229bb30562608af4f41a -> trunk/a6210fd07b8fe1924f24229bb30562608af4f41a 2025-07-17T08:34:52.1895593Z * [new tag] trunk/a666cf3b38f5e7739f94b175f152b127a6d51344 -> trunk/a666cf3b38f5e7739f94b175f152b127a6d51344 2025-07-17T08:34:52.1897812Z * [new tag] trunk/a67eb1a0d627d59d95a99722d37cab3bf054261f -> trunk/a67eb1a0d627d59d95a99722d37cab3bf054261f 2025-07-17T08:34:52.1900071Z * [new tag] trunk/a69e27ca5ad4287add73972ef1b34b469e3c7d23 -> trunk/a69e27ca5ad4287add73972ef1b34b469e3c7d23 2025-07-17T08:34:52.1902566Z * [new tag] trunk/a6a3a441442a96f38d0771c985f753223cea2ba0 -> trunk/a6a3a441442a96f38d0771c985f753223cea2ba0 2025-07-17T08:34:52.1904814Z * [new tag] trunk/a6a8641c8aee5797324c3d41749bcae62b4e5de8 -> trunk/a6a8641c8aee5797324c3d41749bcae62b4e5de8 2025-07-17T08:34:52.1907081Z * [new tag] trunk/a6fab82b16011213cb010c8c50461b9a680748a2 -> trunk/a6fab82b16011213cb010c8c50461b9a680748a2 2025-07-17T08:34:52.1909487Z * [new tag] trunk/a73d9e0aec9319e56ba0c9b0ccc25db69c739faf -> trunk/a73d9e0aec9319e56ba0c9b0ccc25db69c739faf 2025-07-17T08:34:52.1911615Z * [new tag] trunk/a767e50adca9e08dd638be60c9e5604de15b9f15 -> trunk/a767e50adca9e08dd638be60c9e5604de15b9f15 2025-07-17T08:34:52.1913722Z * [new tag] trunk/a7b29c88b1f55bac02f38bf1faba68b6c4be4c9c -> trunk/a7b29c88b1f55bac02f38bf1faba68b6c4be4c9c 2025-07-17T08:34:52.1916218Z * [new tag] trunk/a7eb153bba0534daaa1b652d44bb9fb5bd1ac7c7 -> trunk/a7eb153bba0534daaa1b652d44bb9fb5bd1ac7c7 2025-07-17T08:34:52.1918568Z * [new tag] trunk/a82c171bb26949dda5f5932051a905a9bab38761 -> trunk/a82c171bb26949dda5f5932051a905a9bab38761 2025-07-17T08:34:52.1921069Z * [new tag] trunk/a85ad5552532926b78d489d1b60671f74bf6dc8e -> trunk/a85ad5552532926b78d489d1b60671f74bf6dc8e 2025-07-17T08:34:52.1923173Z * [new tag] trunk/a87dfc7480bc98d33dab274d7e95723aa9cf7840 -> trunk/a87dfc7480bc98d33dab274d7e95723aa9cf7840 2025-07-17T08:34:52.1925465Z * [new tag] trunk/a8b973673798ca79dfe616c9080415d09f9e990d -> trunk/a8b973673798ca79dfe616c9080415d09f9e990d 2025-07-17T08:34:52.1927734Z * [new tag] trunk/a8ec7babcf27584cf068c3a5a04bb3ea0102ab99 -> trunk/a8ec7babcf27584cf068c3a5a04bb3ea0102ab99 2025-07-17T08:34:52.1930019Z * [new tag] trunk/a8fe982993221048ee1665ce28add1b02888784d -> trunk/a8fe982993221048ee1665ce28add1b02888784d 2025-07-17T08:34:52.1932358Z * [new tag] trunk/a92b24cd839aa6807f3c9ea8b8787da99c4490c0 -> trunk/a92b24cd839aa6807f3c9ea8b8787da99c4490c0 2025-07-17T08:34:52.1934953Z * [new tag] trunk/a9352bd25e9f070e8f846b8fdbcd7c90ad5c66b3 -> trunk/a9352bd25e9f070e8f846b8fdbcd7c90ad5c66b3 2025-07-17T08:34:52.1937507Z * [new tag] trunk/a952956d05dd617355007ae31d8e936474a35f14 -> trunk/a952956d05dd617355007ae31d8e936474a35f14 2025-07-17T08:34:52.1948620Z * [new tag] trunk/a9537b626c91ce617139ade60b9107a2805a4248 -> trunk/a9537b626c91ce617139ade60b9107a2805a4248 2025-07-17T08:34:52.1949139Z * [new tag] trunk/a95504b10fff38b5308660e0b535961beed6c9f1 -> trunk/a95504b10fff38b5308660e0b535961beed6c9f1 2025-07-17T08:34:52.1949644Z * [new tag] trunk/a9a0501ec40d8fab07b0f8d2fcbf8eb7147c1ae8 -> trunk/a9a0501ec40d8fab07b0f8d2fcbf8eb7147c1ae8 2025-07-17T08:34:52.1950159Z * [new tag] trunk/a9ac9f2635a9be28eaf267ed41070081a4ce05d8 -> trunk/a9ac9f2635a9be28eaf267ed41070081a4ce05d8 2025-07-17T08:34:52.1950684Z * [new tag] trunk/a9d5157e2596e1d830158e83fab618aab1d4ee39 -> trunk/a9d5157e2596e1d830158e83fab618aab1d4ee39 2025-07-17T08:34:52.1951199Z * [new tag] trunk/a9ee4250d55c6342b80e2d57a8ad9a1992ddcdce -> trunk/a9ee4250d55c6342b80e2d57a8ad9a1992ddcdce 2025-07-17T08:34:52.1953607Z * [new tag] trunk/a9ef7c4d044d386993a6e82b5ad37b7db759b502 -> trunk/a9ef7c4d044d386993a6e82b5ad37b7db759b502 2025-07-17T08:34:52.1956015Z * [new tag] trunk/a9f902add02383ca1b0386eb865767641975fede -> trunk/a9f902add02383ca1b0386eb865767641975fede 2025-07-17T08:34:52.1958426Z * [new tag] trunk/aa116285768ba6a8cd34bf8d245e38d04c88810a -> trunk/aa116285768ba6a8cd34bf8d245e38d04c88810a 2025-07-17T08:34:52.1960641Z * [new tag] trunk/aa280ea19fb20923d048909fa98af092e18ca2fb -> trunk/aa280ea19fb20923d048909fa98af092e18ca2fb 2025-07-17T08:34:52.1962985Z * [new tag] trunk/aa2d54148d476383986855af3fe53862da861dda -> trunk/aa2d54148d476383986855af3fe53862da861dda 2025-07-17T08:34:52.1965239Z * [new tag] trunk/aab949aa96a6b197b75ffa25608fa84049ff68ad -> trunk/aab949aa96a6b197b75ffa25608fa84049ff68ad 2025-07-17T08:34:52.1967578Z * [new tag] trunk/aac0e8f0e998a9381f0eb5dd303a853515109eb0 -> trunk/aac0e8f0e998a9381f0eb5dd303a853515109eb0 2025-07-17T08:34:52.1969820Z * [new tag] trunk/aacb9440791f4e9567e2123f19bb062e75c2d242 -> trunk/aacb9440791f4e9567e2123f19bb062e75c2d242 2025-07-17T08:34:52.1972110Z * [new tag] trunk/ab2294d8289a7757a2fc321cdefac88e2b378edf -> trunk/ab2294d8289a7757a2fc321cdefac88e2b378edf 2025-07-17T08:34:52.1974398Z * [new tag] trunk/ab3393e923514438c9bdfc1afcd0b2aac8be8de9 -> trunk/ab3393e923514438c9bdfc1afcd0b2aac8be8de9 2025-07-17T08:34:52.1976854Z * [new tag] trunk/ab51a937371e61f91cacae7d9fb4b1fe61f24f03 -> trunk/ab51a937371e61f91cacae7d9fb4b1fe61f24f03 2025-07-17T08:34:52.1979161Z * [new tag] trunk/ab56e5add9c36f611a4d9af2f43213926b65d7b4 -> trunk/ab56e5add9c36f611a4d9af2f43213926b65d7b4 2025-07-17T08:34:52.1981384Z * [new tag] trunk/ab655816b8f76f511fb2262d45276d8d1b13d59c -> trunk/ab655816b8f76f511fb2262d45276d8d1b13d59c 2025-07-17T08:34:52.1983843Z * [new tag] trunk/ab6cb34480a14f3cf2446519189ee5f0c5e7278d -> trunk/ab6cb34480a14f3cf2446519189ee5f0c5e7278d 2025-07-17T08:34:52.1985989Z * [new tag] trunk/ab81fb846cd0e52bbc6a2eb366706350639e209f -> trunk/ab81fb846cd0e52bbc6a2eb366706350639e209f 2025-07-17T08:34:52.1988227Z * [new tag] trunk/ab8874bd26894a8ab956408ceee2e89c9ad7f450 -> trunk/ab8874bd26894a8ab956408ceee2e89c9ad7f450 2025-07-17T08:34:52.1990479Z * [new tag] trunk/abbdf9f3632af505b3265b4258a6c185b257edde -> trunk/abbdf9f3632af505b3265b4258a6c185b257edde 2025-07-17T08:34:52.1992871Z * [new tag] trunk/abeae997a35b1920a45be9c26eff7474f2c6c5dd -> trunk/abeae997a35b1920a45be9c26eff7474f2c6c5dd 2025-07-17T08:34:52.1996518Z * [new tag] trunk/abf4da0d242274ba7ab4c0859d105cae74a2dbc5 -> trunk/abf4da0d242274ba7ab4c0859d105cae74a2dbc5 2025-07-17T08:34:52.1998804Z * [new tag] trunk/ac706bfc7f942b8a97401486a840dd8f6452f5cb -> trunk/ac706bfc7f942b8a97401486a840dd8f6452f5cb 2025-07-17T08:34:52.2001097Z * [new tag] trunk/ac86ec0e60370c037e018137f2048cafd47c5c28 -> trunk/ac86ec0e60370c037e018137f2048cafd47c5c28 2025-07-17T08:34:52.2003335Z * [new tag] trunk/acaf6ba3c6d0bdec88ab3f6c2ef82650050558d2 -> trunk/acaf6ba3c6d0bdec88ab3f6c2ef82650050558d2 2025-07-17T08:34:52.2005586Z * [new tag] trunk/ad86c05b78d343f3346d07a07e17639e421d2cf4 -> trunk/ad86c05b78d343f3346d07a07e17639e421d2cf4 2025-07-17T08:34:52.2007895Z * [new tag] trunk/ada44e5ba78be9377814678d1986556af2d6e570 -> trunk/ada44e5ba78be9377814678d1986556af2d6e570 2025-07-17T08:34:52.2010245Z * [new tag] trunk/add0b450bd9907ba9d089c79ca4af96c0590d8ff -> trunk/add0b450bd9907ba9d089c79ca4af96c0590d8ff 2025-07-17T08:34:52.2012558Z * [new tag] trunk/adf6dd1e44c1b006ea6078576b64d5616cecfa5a -> trunk/adf6dd1e44c1b006ea6078576b64d5616cecfa5a 2025-07-17T08:34:52.2015077Z * [new tag] trunk/adf9644440e0d007853e7fdfd67b6a59248f4fbb -> trunk/adf9644440e0d007853e7fdfd67b6a59248f4fbb 2025-07-17T08:34:52.2017393Z * [new tag] trunk/ae0f1f89847564b04bdc43395e664e48f08aaca2 -> trunk/ae0f1f89847564b04bdc43395e664e48f08aaca2 2025-07-17T08:34:52.2019652Z * [new tag] trunk/ae1094b72b7db368f292ed6033de5b495dcdc878 -> trunk/ae1094b72b7db368f292ed6033de5b495dcdc878 2025-07-17T08:34:52.2022055Z * [new tag] trunk/ae86e8f6c829a3cfa9204949156fce2d048c919b -> trunk/ae86e8f6c829a3cfa9204949156fce2d048c919b 2025-07-17T08:34:52.2024336Z * [new tag] trunk/aec569da2325c09487ab659a0f28c3b51e60d779 -> trunk/aec569da2325c09487ab659a0f28c3b51e60d779 2025-07-17T08:34:52.2026631Z * [new tag] trunk/aee2bfc5bae2974f5a3023d51dfa8e3e641144cc -> trunk/aee2bfc5bae2974f5a3023d51dfa8e3e641144cc 2025-07-17T08:34:52.2028855Z * [new tag] trunk/aeffb68d3466635e4e95c50bffd7dfebaba94da2 -> trunk/aeffb68d3466635e4e95c50bffd7dfebaba94da2 2025-07-17T08:34:52.2031218Z * [new tag] trunk/af284b45d5cbfd19d168bf1d118b3a40cd8b3a92 -> trunk/af284b45d5cbfd19d168bf1d118b3a40cd8b3a92 2025-07-17T08:34:52.2033591Z * [new tag] trunk/af3d06909400413801b726a96f3ff52b5ee9ebb5 -> trunk/af3d06909400413801b726a96f3ff52b5ee9ebb5 2025-07-17T08:34:52.2036120Z * [new tag] trunk/af9c92b4cb9f406129dfd8c64c082c0aaf7c723f -> trunk/af9c92b4cb9f406129dfd8c64c082c0aaf7c723f 2025-07-17T08:34:52.2038357Z * [new tag] trunk/aff9c1eec599b8fa97240d192534b0f650c16b1b -> trunk/aff9c1eec599b8fa97240d192534b0f650c16b1b 2025-07-17T08:34:52.2040577Z * [new tag] trunk/b00b641ff1fbfa8f4f6152ffc631e0d24145a7a8 -> trunk/b00b641ff1fbfa8f4f6152ffc631e0d24145a7a8 2025-07-17T08:34:52.2042748Z * [new tag] trunk/b020971e7806bba39aecf636e59e743911831ad8 -> trunk/b020971e7806bba39aecf636e59e743911831ad8 2025-07-17T08:34:52.2045165Z * [new tag] trunk/b0556110e58e3bcf2c872b933e4fd4a0d34398ad -> trunk/b0556110e58e3bcf2c872b933e4fd4a0d34398ad 2025-07-17T08:34:52.2047271Z * [new tag] trunk/b07725a9516028a485153c4b5356b3e33b990f81 -> trunk/b07725a9516028a485153c4b5356b3e33b990f81 2025-07-17T08:34:52.2049613Z * [new tag] trunk/b096341963c4af3b2b0d7f598eba1cdc1c74fb6e -> trunk/b096341963c4af3b2b0d7f598eba1cdc1c74fb6e 2025-07-17T08:34:52.2051924Z * [new tag] trunk/b09bd414a6ccba158c09f586a278051588d90936 -> trunk/b09bd414a6ccba158c09f586a278051588d90936 2025-07-17T08:34:52.2054221Z * [new tag] trunk/b0fbbef1361ccaab8a5aec8e7cd62150e7b361de -> trunk/b0fbbef1361ccaab8a5aec8e7cd62150e7b361de 2025-07-17T08:34:52.2056453Z * [new tag] trunk/b146ca74f01df3cf711fd0f855e05805e490156c -> trunk/b146ca74f01df3cf711fd0f855e05805e490156c 2025-07-17T08:34:52.2058664Z * [new tag] trunk/b146e1a264960e5fb79cf4621ebe4b1345db3e64 -> trunk/b146e1a264960e5fb79cf4621ebe4b1345db3e64 2025-07-17T08:34:52.2060949Z * [new tag] trunk/b147b6c0e31d43a405416743fde1b0b606aa3839 -> trunk/b147b6c0e31d43a405416743fde1b0b606aa3839 2025-07-17T08:34:52.2063163Z * [new tag] trunk/b1713c665516d8355f5491d10b31f98c74ce27f1 -> trunk/b1713c665516d8355f5491d10b31f98c74ce27f1 2025-07-17T08:34:52.2065481Z * [new tag] trunk/b1a54fab9bcb0cc167773f9a885d4170447e1c68 -> trunk/b1a54fab9bcb0cc167773f9a885d4170447e1c68 2025-07-17T08:34:52.2067698Z * [new tag] trunk/b1b8e57cda03bd016594a4fc4944fa5e7f9dd961 -> trunk/b1b8e57cda03bd016594a4fc4944fa5e7f9dd961 2025-07-17T08:34:52.2070049Z * [new tag] trunk/b1d62febd03ac421197d5516596f98d3c46e9b44 -> trunk/b1d62febd03ac421197d5516596f98d3c46e9b44 2025-07-17T08:34:52.2072316Z * [new tag] trunk/b221be9140689d180ec339db05c5e235a95949d3 -> trunk/b221be9140689d180ec339db05c5e235a95949d3 2025-07-17T08:34:52.2074856Z * [new tag] trunk/b22b93a6babf86e823e979528bc0aab88fb7a012 -> trunk/b22b93a6babf86e823e979528bc0aab88fb7a012 2025-07-17T08:34:52.2077174Z * [new tag] trunk/b26da7741be37693ab1cd21115f3fca15b1cdb6b -> trunk/b26da7741be37693ab1cd21115f3fca15b1cdb6b 2025-07-17T08:34:52.2079419Z * [new tag] trunk/b2d473c8f8a6d0677940c174b38511f5ab3c3b65 -> trunk/b2d473c8f8a6d0677940c174b38511f5ab3c3b65 2025-07-17T08:34:52.2081812Z * [new tag] trunk/b2fc9cfea16c8eb52c1ce79b2032793dd1a545fb -> trunk/b2fc9cfea16c8eb52c1ce79b2032793dd1a545fb 2025-07-17T08:34:52.2084061Z * [new tag] trunk/b30e04b3c8a83e19dc9123f2f0229d80e208fb50 -> trunk/b30e04b3c8a83e19dc9123f2f0229d80e208fb50 2025-07-17T08:34:52.2086352Z * [new tag] trunk/b354328ecd6a313eab905ab6f74aed5ea2d3a2ce -> trunk/b354328ecd6a313eab905ab6f74aed5ea2d3a2ce 2025-07-17T08:34:52.2088709Z * [new tag] trunk/b359571c6043b40c4ae4fbb07135fd0f04902e21 -> trunk/b359571c6043b40c4ae4fbb07135fd0f04902e21 2025-07-17T08:34:52.2091001Z * [new tag] trunk/b3b4d28f4cf149cb7448c556d359ca852d815ab1 -> trunk/b3b4d28f4cf149cb7448c556d359ca852d815ab1 2025-07-17T08:34:52.2093258Z * [new tag] trunk/b40981c63078d99bb07afbdcfec1ce8f7519f26b -> trunk/b40981c63078d99bb07afbdcfec1ce8f7519f26b 2025-07-17T08:34:52.2095665Z * [new tag] trunk/b40c0b61eb5f5b6104cb732e709819e84518faa7 -> trunk/b40c0b61eb5f5b6104cb732e709819e84518faa7 2025-07-17T08:34:52.2097864Z * [new tag] trunk/b4228a94d11b1ba6599f443267824d2d918644f2 -> trunk/b4228a94d11b1ba6599f443267824d2d918644f2 2025-07-17T08:34:52.2100366Z * [new tag] trunk/b44306d3681d5b248e6b439d293ea0d5a8903a61 -> trunk/b44306d3681d5b248e6b439d293ea0d5a8903a61 2025-07-17T08:34:52.2102634Z * [new tag] trunk/b4442f42a93390760bb923cbe13b80993f5e8e78 -> trunk/b4442f42a93390760bb923cbe13b80993f5e8e78 2025-07-17T08:34:52.2104831Z * [new tag] trunk/b4476ca378be50034bd5cdc1eaa95104337c998a -> trunk/b4476ca378be50034bd5cdc1eaa95104337c998a 2025-07-17T08:34:52.2107506Z * [new tag] trunk/b487003182b7f2d6697064f184515369bf6c8cce -> trunk/b487003182b7f2d6697064f184515369bf6c8cce 2025-07-17T08:34:52.2109552Z * [new tag] trunk/b49edc0d6c72f744d744a026b0ce0d689e34b83b -> trunk/b49edc0d6c72f744d744a026b0ce0d689e34b83b 2025-07-17T08:34:52.2111792Z * [new tag] trunk/b4e3c9ea34cb607324639cd0bb0129740c300721 -> trunk/b4e3c9ea34cb607324639cd0bb0129740c300721 2025-07-17T08:34:52.2114152Z * [new tag] trunk/b4fc42ca807efc1a1b78f52d5481711a4e8bbea3 -> trunk/b4fc42ca807efc1a1b78f52d5481711a4e8bbea3 2025-07-17T08:34:52.2116517Z * [new tag] trunk/b50075343aeb519d135e00c44ff3577eaa25c61b -> trunk/b50075343aeb519d135e00c44ff3577eaa25c61b 2025-07-17T08:34:52.2118798Z * [new tag] trunk/b54eac2a5ed31106393bbc338de8637817809a1f -> trunk/b54eac2a5ed31106393bbc338de8637817809a1f 2025-07-17T08:34:52.2121131Z * [new tag] trunk/b5bfbba1841da810305262c7f47ee2dae54f335e -> trunk/b5bfbba1841da810305262c7f47ee2dae54f335e 2025-07-17T08:34:52.2123415Z * [new tag] trunk/b5c8b8d09f006b1b2911858882a56dfe6e325f36 -> trunk/b5c8b8d09f006b1b2911858882a56dfe6e325f36 2025-07-17T08:34:52.2125733Z * [new tag] trunk/b5ce77c1f5964293299eb1366f341872a4e47fa6 -> trunk/b5ce77c1f5964293299eb1366f341872a4e47fa6 2025-07-17T08:34:52.2128096Z * [new tag] trunk/b60569ed946ddcc267150a675916d68d7cac9085 -> trunk/b60569ed946ddcc267150a675916d68d7cac9085 2025-07-17T08:34:52.2130387Z * [new tag] trunk/b6188174795956f959feae0cbc33cbdb7901c4b6 -> trunk/b6188174795956f959feae0cbc33cbdb7901c4b6 2025-07-17T08:34:52.2132624Z * [new tag] trunk/b642a5c118baf4cd47dd7fe0190b83e04cee960f -> trunk/b642a5c118baf4cd47dd7fe0190b83e04cee960f 2025-07-17T08:34:52.2134888Z * [new tag] trunk/b6454a9058f2e50be9a3c26c128fec843b09c154 -> trunk/b6454a9058f2e50be9a3c26c128fec843b09c154 2025-07-17T08:34:52.2137274Z * [new tag] trunk/b6add8c8ba927a9f1687134ec2734b02a6ccb3c0 -> trunk/b6add8c8ba927a9f1687134ec2734b02a6ccb3c0 2025-07-17T08:34:52.2139614Z * [new tag] trunk/b6c00dfe249a7bfc1d61a322d5bc30f164353abf -> trunk/b6c00dfe249a7bfc1d61a322d5bc30f164353abf 2025-07-17T08:34:52.2141985Z * [new tag] trunk/b6e625e34f358c71b62409f96dc1e22e4791beef -> trunk/b6e625e34f358c71b62409f96dc1e22e4791beef 2025-07-17T08:34:52.2144427Z * [new tag] trunk/b6f84b3b0fef781653911420253dcff6767197dc -> trunk/b6f84b3b0fef781653911420253dcff6767197dc 2025-07-17T08:34:52.2146722Z * [new tag] trunk/b754b1fa43d20f5b31e17c396487ab56991912da -> trunk/b754b1fa43d20f5b31e17c396487ab56991912da 2025-07-17T08:34:52.2149095Z * [new tag] trunk/b7860c7863df9139bceabf5b0c186e6e4287aa4c -> trunk/b7860c7863df9139bceabf5b0c186e6e4287aa4c 2025-07-17T08:34:52.2151414Z * [new tag] trunk/b7a73a2cdbbe46e04aa145fb8c1615e608dcbde9 -> trunk/b7a73a2cdbbe46e04aa145fb8c1615e608dcbde9 2025-07-17T08:34:52.2153677Z * [new tag] trunk/b7b1109f49f5d0bd6145ae47c5c7d7d18c5659b0 -> trunk/b7b1109f49f5d0bd6145ae47c5c7d7d18c5659b0 2025-07-17T08:34:52.2156292Z * [new tag] trunk/b7def5ff1ca72fbb06350ffc75117efc68e149fb -> trunk/b7def5ff1ca72fbb06350ffc75117efc68e149fb 2025-07-17T08:34:52.2158557Z * [new tag] trunk/b83d8827bcd63501d7298267d94d103bf591c6c2 -> trunk/b83d8827bcd63501d7298267d94d103bf591c6c2 2025-07-17T08:34:52.2160864Z * [new tag] trunk/b85f10ea5006e8ae8fc769f48659ab7ad5eafb69 -> trunk/b85f10ea5006e8ae8fc769f48659ab7ad5eafb69 2025-07-17T08:34:52.2163203Z * [new tag] trunk/b86d5cef68d56f3924dc199424e65904a32d0743 -> trunk/b86d5cef68d56f3924dc199424e65904a32d0743 2025-07-17T08:34:52.2165647Z * [new tag] trunk/b878ca0c91bdccbdd907c15a01e5dcf249f0783c -> trunk/b878ca0c91bdccbdd907c15a01e5dcf249f0783c 2025-07-17T08:34:52.2168175Z * [new tag] trunk/b8ace6f95105751d0535cb0ce0d2c8f656c8b96c -> trunk/b8ace6f95105751d0535cb0ce0d2c8f656c8b96c 2025-07-17T08:34:52.2170374Z * [new tag] trunk/b8aee84fb9692f1a46c8c2a0ae6685df440f6e8f -> trunk/b8aee84fb9692f1a46c8c2a0ae6685df440f6e8f 2025-07-17T08:34:52.2172615Z * [new tag] trunk/b8bc2c2660e84034ff15232e2161e3ef9a6656d0 -> trunk/b8bc2c2660e84034ff15232e2161e3ef9a6656d0 2025-07-17T08:34:52.2174865Z * [new tag] trunk/b8c2d4c259b16f3dc044d049d964c44907ccecb6 -> trunk/b8c2d4c259b16f3dc044d049d964c44907ccecb6 2025-07-17T08:34:52.2177151Z * [new tag] trunk/b8d96c3f78a27e193f4fa9580f8d28298c8180e3 -> trunk/b8d96c3f78a27e193f4fa9580f8d28298c8180e3 2025-07-17T08:34:52.2179410Z * [new tag] trunk/b8fc5e0c0dda5da8315f9bc108a83a0f92252513 -> trunk/b8fc5e0c0dda5da8315f9bc108a83a0f92252513 2025-07-17T08:34:52.2181749Z * [new tag] trunk/b916d8a583cf4fb018c44eaf5efefa9ca76da366 -> trunk/b916d8a583cf4fb018c44eaf5efefa9ca76da366 2025-07-17T08:34:52.2184014Z * [new tag] trunk/b95dadd7170626273bb03b1264f04f3f051908da -> trunk/b95dadd7170626273bb03b1264f04f3f051908da 2025-07-17T08:34:52.2186253Z * [new tag] trunk/b981fb6744d034a0b2a0baf7e138cfb8dc83ca96 -> trunk/b981fb6744d034a0b2a0baf7e138cfb8dc83ca96 2025-07-17T08:34:52.2188587Z * [new tag] trunk/b9afdd9bcc738697c6eefc90899508ab783bf6ab -> trunk/b9afdd9bcc738697c6eefc90899508ab783bf6ab 2025-07-17T08:34:52.2190985Z * [new tag] trunk/b9b84d8011b08ac62cabf9100043c65863372fea -> trunk/b9b84d8011b08ac62cabf9100043c65863372fea 2025-07-17T08:34:52.2193298Z * [new tag] trunk/b9dc2fa4f7aa237a19248705abf82f5eae662182 -> trunk/b9dc2fa4f7aa237a19248705abf82f5eae662182 2025-07-17T08:34:52.2195799Z * [new tag] trunk/ba0d0de5e652650f8a59c85fc25a90f4ed9e2dc1 -> trunk/ba0d0de5e652650f8a59c85fc25a90f4ed9e2dc1 2025-07-17T08:34:52.2198116Z * [new tag] trunk/bb1f3d1a55e3f8eafb279995824c1ef87f24b341 -> trunk/bb1f3d1a55e3f8eafb279995824c1ef87f24b341 2025-07-17T08:34:52.2200471Z * [new tag] trunk/bb3c911c2d64ccad03349e50a0a220a2889ae85c -> trunk/bb3c911c2d64ccad03349e50a0a220a2889ae85c 2025-07-17T08:34:52.2202887Z * [new tag] trunk/bb462a6237c163774c99e01462703ebce55f4589 -> trunk/bb462a6237c163774c99e01462703ebce55f4589 2025-07-17T08:34:52.2205111Z * [new tag] trunk/bb476310a456b1fd418c79573ed34ad68b4871d4 -> trunk/bb476310a456b1fd418c79573ed34ad68b4871d4 2025-07-17T08:34:52.2207435Z * [new tag] trunk/bbb930aba2e769e3ed981f2a035133891de36dae -> trunk/bbb930aba2e769e3ed981f2a035133891de36dae 2025-07-17T08:34:52.2209692Z * [new tag] trunk/bbbced94a43cf764ddfe719e7d4c161a3992830c -> trunk/bbbced94a43cf764ddfe719e7d4c161a3992830c 2025-07-17T08:34:52.2212058Z * [new tag] trunk/bbe681ed510227ce05f99db2a36c1af5c2cc302b -> trunk/bbe681ed510227ce05f99db2a36c1af5c2cc302b 2025-07-17T08:34:52.2214266Z * [new tag] trunk/bbf1a6feac0e88772be103a1b159b871b5c00b4a -> trunk/bbf1a6feac0e88772be103a1b159b871b5c00b4a 2025-07-17T08:34:52.2216646Z * [new tag] trunk/bc3972b80a7abe85036f48b610532fce39ea5097 -> trunk/bc3972b80a7abe85036f48b610532fce39ea5097 2025-07-17T08:34:52.2218967Z * [new tag] trunk/bc5a11b58180d40157175f45c69d60a9b9961315 -> trunk/bc5a11b58180d40157175f45c69d60a9b9961315 2025-07-17T08:34:52.2221247Z * [new tag] trunk/bc65253369933160a2da3fc786d027a572faf6b7 -> trunk/bc65253369933160a2da3fc786d027a572faf6b7 2025-07-17T08:34:52.2223498Z * [new tag] trunk/bc6e0661a6ec7e536bee60b9c929f71643bb6c89 -> trunk/bc6e0661a6ec7e536bee60b9c929f71643bb6c89 2025-07-17T08:34:52.2225838Z * [new tag] trunk/bc9091a524a1ebe4de16af4dd8f442db7d1cb138 -> trunk/bc9091a524a1ebe4de16af4dd8f442db7d1cb138 2025-07-17T08:34:52.2228122Z * [new tag] trunk/bc9bd2a766ae8cf2aba8f3d0153cd8ecd05b4465 -> trunk/bc9bd2a766ae8cf2aba8f3d0153cd8ecd05b4465 2025-07-17T08:34:52.2230558Z * [new tag] trunk/bcad962550f2bfd850684250a3c881c9f38ad601 -> trunk/bcad962550f2bfd850684250a3c881c9f38ad601 2025-07-17T08:34:52.2232673Z * [new tag] trunk/bcc98bb2a4dd11ac696082731f8980b72deb6750 -> trunk/bcc98bb2a4dd11ac696082731f8980b72deb6750 2025-07-17T08:34:52.2237162Z * [new tag] trunk/bccb8473fed94dbc6f1392d0c5b4a51150ee4a9a -> trunk/bccb8473fed94dbc6f1392d0c5b4a51150ee4a9a 2025-07-17T08:34:52.2239543Z * [new tag] trunk/bcf50636ba1b93a833267c645d887888df06e9ea -> trunk/bcf50636ba1b93a833267c645d887888df06e9ea 2025-07-17T08:34:52.2241799Z * [new tag] trunk/bd364c901d5b20500ec5cbe275e93c955809d900 -> trunk/bd364c901d5b20500ec5cbe275e93c955809d900 2025-07-17T08:34:52.2244027Z * [new tag] trunk/bd3c32916cf5f85ddb9dd3c36e7311adfa8808af -> trunk/bd3c32916cf5f85ddb9dd3c36e7311adfa8808af 2025-07-17T08:34:52.2246421Z * [new tag] trunk/bd6b5fddbf5fb0b603ab8a7428079d9a86cf532a -> trunk/bd6b5fddbf5fb0b603ab8a7428079d9a86cf532a 2025-07-17T08:34:52.2248656Z * [new tag] trunk/bdacf08b8682b9fbe3a8656a53b1f8b1cb007fd8 -> trunk/bdacf08b8682b9fbe3a8656a53b1f8b1cb007fd8 2025-07-17T08:34:52.2251102Z * [new tag] trunk/bdb1553b77bb28df580ea41d726417ac91028ec6 -> trunk/bdb1553b77bb28df580ea41d726417ac91028ec6 2025-07-17T08:34:52.2253368Z * [new tag] trunk/bdb78191662c01ab1263108febac29a8560337d0 -> trunk/bdb78191662c01ab1263108febac29a8560337d0 2025-07-17T08:34:52.2255644Z * [new tag] trunk/bdbf2792a8f15907f6bbf231f73dab0e8efe1c50 -> trunk/bdbf2792a8f15907f6bbf231f73dab0e8efe1c50 2025-07-17T08:34:52.2258023Z * [new tag] trunk/be124a61a4933603795644d068d2c0e5f444e766 -> trunk/be124a61a4933603795644d068d2c0e5f444e766 2025-07-17T08:34:52.2260431Z * [new tag] trunk/be2ab96347a1e7206a57f69b58263c4455ff8f76 -> trunk/be2ab96347a1e7206a57f69b58263c4455ff8f76 2025-07-17T08:34:52.2262835Z * [new tag] trunk/be2ad70cfa1360da5c23a04ff6ca3480fa02f278 -> trunk/be2ad70cfa1360da5c23a04ff6ca3480fa02f278 2025-07-17T08:34:52.2265149Z * [new tag] trunk/be2e43264d0bce254efabcc60c368d1418bde57f -> trunk/be2e43264d0bce254efabcc60c368d1418bde57f 2025-07-17T08:34:52.2267457Z * [new tag] trunk/be56a8d7ac0dc72f6946354d7ac5be9ece0f1c35 -> trunk/be56a8d7ac0dc72f6946354d7ac5be9ece0f1c35 2025-07-17T08:34:52.2269733Z * [new tag] trunk/beb52f5c0ac8efddeafa862f274bc247db989695 -> trunk/beb52f5c0ac8efddeafa862f274bc247db989695 2025-07-17T08:34:52.2272032Z * [new tag] trunk/bee93f9f0d16f4f563812bb5c16e862de15724c1 -> trunk/bee93f9f0d16f4f563812bb5c16e862de15724c1 2025-07-17T08:34:52.2274221Z * [new tag] trunk/beed033b6e6ac57c0b4a1f47eb436e115a52e41b -> trunk/beed033b6e6ac57c0b4a1f47eb436e115a52e41b 2025-07-17T08:34:52.2276891Z * [new tag] trunk/bf06190e21fdf539c13e1ec01271653d0729733a -> trunk/bf06190e21fdf539c13e1ec01271653d0729733a 2025-07-17T08:34:52.2279198Z * [new tag] trunk/bf1ebe0531e1b0390ea09d66ac78500d3c6d3a15 -> trunk/bf1ebe0531e1b0390ea09d66ac78500d3c6d3a15 2025-07-17T08:34:52.2281415Z * [new tag] trunk/bf50d715539acedcb31a6d8f787149eff6213fdb -> trunk/bf50d715539acedcb31a6d8f787149eff6213fdb 2025-07-17T08:34:52.2283707Z * [new tag] trunk/bf798a2f016ca0001750436150e7a2bdb2676d1a -> trunk/bf798a2f016ca0001750436150e7a2bdb2676d1a 2025-07-17T08:34:52.2285827Z * [new tag] trunk/bf7e290854b7f0ab3fb89251d0100821f1b3a70e -> trunk/bf7e290854b7f0ab3fb89251d0100821f1b3a70e 2025-07-17T08:34:52.2288500Z * [new tag] trunk/bf897b4cea2b7481499a81afa30bfc69a8e685c4 -> trunk/bf897b4cea2b7481499a81afa30bfc69a8e685c4 2025-07-17T08:34:52.2290727Z * [new tag] trunk/bfcababbcb95ec42046737dbcf61f8e258075ace -> trunk/bfcababbcb95ec42046737dbcf61f8e258075ace 2025-07-17T08:34:52.2293141Z * [new tag] trunk/bfccfa0b31221d5df0b263de5a41fb9f7c84b97d -> trunk/bfccfa0b31221d5df0b263de5a41fb9f7c84b97d 2025-07-17T08:34:52.2295209Z * [new tag] trunk/c038719731abdcd415a5e82c5f3826f2358229fc -> trunk/c038719731abdcd415a5e82c5f3826f2358229fc 2025-07-17T08:34:52.2297473Z * [new tag] trunk/c04a4e709410936c12cdc9db5aaf47164f74a367 -> trunk/c04a4e709410936c12cdc9db5aaf47164f74a367 2025-07-17T08:34:52.2299980Z * [new tag] trunk/c062550a3598d27c2d6572db7c0f4ff90a84cc84 -> trunk/c062550a3598d27c2d6572db7c0f4ff90a84cc84 2025-07-17T08:34:52.2302378Z * [new tag] trunk/c06c2569ee53c169c96f42b45bdb64f390d2f23b -> trunk/c06c2569ee53c169c96f42b45bdb64f390d2f23b 2025-07-17T08:34:52.2304673Z * [new tag] trunk/c09b05487877a3f8b239523b3806f5dd9fc82051 -> trunk/c09b05487877a3f8b239523b3806f5dd9fc82051 2025-07-17T08:34:52.2306955Z * [new tag] trunk/c09cf29d7d010c0547f2f6771921a0ee2ec4904b -> trunk/c09cf29d7d010c0547f2f6771921a0ee2ec4904b 2025-07-17T08:34:52.2309290Z * [new tag] trunk/c09eba877f9c16908b3a925ef694604c1c761b85 -> trunk/c09eba877f9c16908b3a925ef694604c1c761b85 2025-07-17T08:34:52.2311580Z * [new tag] trunk/c0e155a8d230b04f186d1b7540b01213be2b9fc6 -> trunk/c0e155a8d230b04f186d1b7540b01213be2b9fc6 2025-07-17T08:34:52.2314553Z * [new tag] trunk/c0ee01c2fb54e2ae924270f7fd53069915a61a50 -> trunk/c0ee01c2fb54e2ae924270f7fd53069915a61a50 2025-07-17T08:34:52.2316934Z * [new tag] trunk/c10339559de1464c2c5aeb16649c382eefcfb572 -> trunk/c10339559de1464c2c5aeb16649c382eefcfb572 2025-07-17T08:34:52.2319194Z * [new tag] trunk/c11888e7a6557ff64e24ca06e22946c116a7c1e4 -> trunk/c11888e7a6557ff64e24ca06e22946c116a7c1e4 2025-07-17T08:34:52.2321630Z * [new tag] trunk/c13e725edd8dd21406c629bf625f2d6c59ceedd1 -> trunk/c13e725edd8dd21406c629bf625f2d6c59ceedd1 2025-07-17T08:34:52.2323979Z * [new tag] trunk/c14110056f1b6e989850f20ae56d47fbc775e890 -> trunk/c14110056f1b6e989850f20ae56d47fbc775e890 2025-07-17T08:34:52.2326214Z * [new tag] trunk/c1446e1e9d2f5a1bae5fb3d58e58f16c15bb15c3 -> trunk/c1446e1e9d2f5a1bae5fb3d58e58f16c15bb15c3 2025-07-17T08:34:52.2328530Z * [new tag] trunk/c165b36a31585d4e061a84e1977428afd931b82d -> trunk/c165b36a31585d4e061a84e1977428afd931b82d 2025-07-17T08:34:52.2330652Z * [new tag] trunk/c174f3a6a55864cedb8f6d9014e9b8cadf91186b -> trunk/c174f3a6a55864cedb8f6d9014e9b8cadf91186b 2025-07-17T08:34:52.2333053Z * [new tag] trunk/c177abd217ecef5cc096d50adbf5990525714dd3 -> trunk/c177abd217ecef5cc096d50adbf5990525714dd3 2025-07-17T08:34:52.2335413Z * [new tag] trunk/c199a4d0fd03ab3f7e36c6133839cfaf34ba59da -> trunk/c199a4d0fd03ab3f7e36c6133839cfaf34ba59da 2025-07-17T08:34:52.2337668Z * [new tag] trunk/c1a629f76d0afc2ddb4a5c7c1925caa1d12b5a5e -> trunk/c1a629f76d0afc2ddb4a5c7c1925caa1d12b5a5e 2025-07-17T08:34:52.2339903Z * [new tag] trunk/c1ad4b8e7a16f54c35a3908b56ed7d9f95eef586 -> trunk/c1ad4b8e7a16f54c35a3908b56ed7d9f95eef586 2025-07-17T08:34:52.2342178Z * [new tag] trunk/c1ae768baa9598ea97f301999a227b34a2efdecf -> trunk/c1ae768baa9598ea97f301999a227b34a2efdecf 2025-07-17T08:34:52.2344450Z * [new tag] trunk/c1cbaca7fd9937c0a089d98aa69066daf35c898f -> trunk/c1cbaca7fd9937c0a089d98aa69066daf35c898f 2025-07-17T08:34:52.2346709Z * [new tag] trunk/c1f531f0b0e6faf443d90f8de2936e866c8c27c2 -> trunk/c1f531f0b0e6faf443d90f8de2936e866c8c27c2 2025-07-17T08:34:52.2349016Z * [new tag] trunk/c202a7329ad798da676762a8af2aa588f882d288 -> trunk/c202a7329ad798da676762a8af2aa588f882d288 2025-07-17T08:34:52.2351465Z * [new tag] trunk/c2185dc4a5626848df37cad214b73d5ae7dd4f17 -> trunk/c2185dc4a5626848df37cad214b73d5ae7dd4f17 2025-07-17T08:34:52.2353962Z * [new tag] trunk/c219dbd2fc70227ba543c98e9740a84723ba36da -> trunk/c219dbd2fc70227ba543c98e9740a84723ba36da 2025-07-17T08:34:52.2356073Z * [new tag] trunk/c2510fcd86152028c3e6cf483740b177a10ac9b9 -> trunk/c2510fcd86152028c3e6cf483740b177a10ac9b9 2025-07-17T08:34:52.2358345Z * [new tag] trunk/c26ce593d8737cab4b55483b62956a6fea9e4375 -> trunk/c26ce593d8737cab4b55483b62956a6fea9e4375 2025-07-17T08:34:52.2360592Z * [new tag] trunk/c27f83dd91ff3599b5157866e6be014c20b967a0 -> trunk/c27f83dd91ff3599b5157866e6be014c20b967a0 2025-07-17T08:34:52.2362916Z * [new tag] trunk/c28e74e45743feac10559c30dbf71cc35bc12ccb -> trunk/c28e74e45743feac10559c30dbf71cc35bc12ccb 2025-07-17T08:34:52.2365213Z * [new tag] trunk/c2beeadeb40e927c51bc5bcd409b3f28374a5190 -> trunk/c2beeadeb40e927c51bc5bcd409b3f28374a5190 2025-07-17T08:34:52.2367559Z * [new tag] trunk/c2d1b225e62bdf3adbba91891f53bc60315adaac -> trunk/c2d1b225e62bdf3adbba91891f53bc60315adaac 2025-07-17T08:34:52.2369925Z * [new tag] trunk/c2f4cc59a70bfd7d7e46e9b5596bae8e4ae2cf9c -> trunk/c2f4cc59a70bfd7d7e46e9b5596bae8e4ae2cf9c 2025-07-17T08:34:52.2372258Z * [new tag] trunk/c329a8f19cc94cb7f9d3fc814484cded138ee3ca -> trunk/c329a8f19cc94cb7f9d3fc814484cded138ee3ca 2025-07-17T08:34:52.2374478Z * [new tag] trunk/c37ddcaefbe9b877e1816ce97dedb8ad26d09450 -> trunk/c37ddcaefbe9b877e1816ce97dedb8ad26d09450 2025-07-17T08:34:52.2376740Z * [new tag] trunk/c3ecabf0593066750156989fc75acbeeaddf0791 -> trunk/c3ecabf0593066750156989fc75acbeeaddf0791 2025-07-17T08:34:52.2379070Z * [new tag] trunk/c48d0f4643b7a69ebe24069e932ce1465a31cdbe -> trunk/c48d0f4643b7a69ebe24069e932ce1465a31cdbe 2025-07-17T08:34:52.2381350Z * [new tag] trunk/c4b93e6579c8e2f1252067c4923cbb859c476fb4 -> trunk/c4b93e6579c8e2f1252067c4923cbb859c476fb4 2025-07-17T08:34:52.2383655Z * [new tag] trunk/c4cdcda754e2b73b030902dd06fb651c7a026510 -> trunk/c4cdcda754e2b73b030902dd06fb651c7a026510 2025-07-17T08:34:52.2385901Z * [new tag] trunk/c515385b0ac4a94deef652159e71fe0912615d14 -> trunk/c515385b0ac4a94deef652159e71fe0912615d14 2025-07-17T08:34:52.2388389Z * [new tag] trunk/c54778625ec45b94eb97f52da3ad6c2eb3d3c3c9 -> trunk/c54778625ec45b94eb97f52da3ad6c2eb3d3c3c9 2025-07-17T08:34:52.2390665Z * [new tag] trunk/c553c55be76daaf70d977a4ac7664b46cbcfa7c4 -> trunk/c553c55be76daaf70d977a4ac7664b46cbcfa7c4 2025-07-17T08:34:52.2392968Z * [new tag] trunk/c5589074e64b28afd737174ae1e9c2ba0b925e56 -> trunk/c5589074e64b28afd737174ae1e9c2ba0b925e56 2025-07-17T08:34:52.2395520Z * [new tag] trunk/c55eef79f8880e4b610a0ca6f6131e690dc948dd -> trunk/c55eef79f8880e4b610a0ca6f6131e690dc948dd 2025-07-17T08:34:52.2397836Z * [new tag] trunk/c5a4fe9c17bc20eb46b15fda554d17030bb6a5b9 -> trunk/c5a4fe9c17bc20eb46b15fda554d17030bb6a5b9 2025-07-17T08:34:52.2400113Z * [new tag] trunk/c5b46b5408e136a5a3060b11a96afa51bd009fd5 -> trunk/c5b46b5408e136a5a3060b11a96afa51bd009fd5 2025-07-17T08:34:52.2402424Z * [new tag] trunk/c5d00e150ad503510bf28adf09f69201aeb94f5c -> trunk/c5d00e150ad503510bf28adf09f69201aeb94f5c 2025-07-17T08:34:52.2404637Z * [new tag] trunk/c5d3e7a4ff460eed70b8443485a7e3568e87aee9 -> trunk/c5d3e7a4ff460eed70b8443485a7e3568e87aee9 2025-07-17T08:34:52.2407003Z * [new tag] trunk/c60327ba74f4db232a2832f2c7ca4b2db43a3132 -> trunk/c60327ba74f4db232a2832f2c7ca4b2db43a3132 2025-07-17T08:34:52.2409363Z * [new tag] trunk/c60d8188d23801d44343c96746ec2e7d2971c5d7 -> trunk/c60d8188d23801d44343c96746ec2e7d2971c5d7 2025-07-17T08:34:52.2411704Z * [new tag] trunk/c620d0b5c7e8679413d620624725471223ce8359 -> trunk/c620d0b5c7e8679413d620624725471223ce8359 2025-07-17T08:34:52.2413875Z * [new tag] trunk/c68af9af1b3652a8e25bd6d0ff8dae89f206a81a -> trunk/c68af9af1b3652a8e25bd6d0ff8dae89f206a81a 2025-07-17T08:34:52.2416290Z * [new tag] trunk/c6a27bae3656021720074520ac6dad9cebe83ff5 -> trunk/c6a27bae3656021720074520ac6dad9cebe83ff5 2025-07-17T08:34:52.2419331Z * [new tag] trunk/c6b4f98625bb6b22bb9a60112a6d58e684a97e1b -> trunk/c6b4f98625bb6b22bb9a60112a6d58e684a97e1b 2025-07-17T08:34:52.2421804Z * [new tag] trunk/c74fd35050a7241f0c439501ef735aa6cdde751f -> trunk/c74fd35050a7241f0c439501ef735aa6cdde751f 2025-07-17T08:34:52.2424125Z * [new tag] trunk/c75c732481915c25295a0ae6321d39aa314eaa93 -> trunk/c75c732481915c25295a0ae6321d39aa314eaa93 2025-07-17T08:34:52.2426504Z * [new tag] trunk/c78bbdf4102d2c13bf6aa1abe4352aa7bca401ca -> trunk/c78bbdf4102d2c13bf6aa1abe4352aa7bca401ca 2025-07-17T08:34:52.2428802Z * [new tag] trunk/c78fce9e79b79686b87f4007cbaec819bdd0223f -> trunk/c78fce9e79b79686b87f4007cbaec819bdd0223f 2025-07-17T08:34:52.2431076Z * [new tag] trunk/c79c7bbe615265b6b3d7df39d6d5a68afd7d6b2a -> trunk/c79c7bbe615265b6b3d7df39d6d5a68afd7d6b2a 2025-07-17T08:34:52.2433235Z * [new tag] trunk/c7b6c98d1097bec9dc92bde2fe324aa126a5daa2 -> trunk/c7b6c98d1097bec9dc92bde2fe324aa126a5daa2 2025-07-17T08:34:52.2435687Z * [new tag] trunk/c808af514d59e25ea4a880c1e1e07d3232984e5d -> trunk/c808af514d59e25ea4a880c1e1e07d3232984e5d 2025-07-17T08:34:52.2437946Z * [new tag] trunk/c811f41cf58238b49e9c2f0a9fa00dcc471b2e10 -> trunk/c811f41cf58238b49e9c2f0a9fa00dcc471b2e10 2025-07-17T08:34:52.2440211Z * [new tag] trunk/c82a174ceab79f77ba18405dd263eb03692608fd -> trunk/c82a174ceab79f77ba18405dd263eb03692608fd 2025-07-17T08:34:52.2442498Z * [new tag] trunk/c83041cac2e9ca537b210b74876dbcee5e08ae9a -> trunk/c83041cac2e9ca537b210b74876dbcee5e08ae9a 2025-07-17T08:34:52.2445355Z * [new tag] trunk/c843909d9e32f92b2e31cf9b8f066daf311a6f18 -> trunk/c843909d9e32f92b2e31cf9b8f066daf311a6f18 2025-07-17T08:34:52.2447655Z * [new tag] trunk/c88e87f3553b7ef0b743fb030317fabb6fe2dc29 -> trunk/c88e87f3553b7ef0b743fb030317fabb6fe2dc29 2025-07-17T08:34:52.2449909Z * [new tag] trunk/c8c221c0b3abbb8b5e20138080644dd5f5cd0aa1 -> trunk/c8c221c0b3abbb8b5e20138080644dd5f5cd0aa1 2025-07-17T08:34:52.2452158Z * [new tag] trunk/c8c892b4a50fdad624753db5b8d2bbb0fc5bb110 -> trunk/c8c892b4a50fdad624753db5b8d2bbb0fc5bb110 2025-07-17T08:34:52.2454397Z * [new tag] trunk/c8d39a10457ea5d65184c6e8f037f46c5525d869 -> trunk/c8d39a10457ea5d65184c6e8f037f46c5525d869 2025-07-17T08:34:52.2456721Z * [new tag] trunk/c9174a20f75ba007b0f520326ced5c65d8a2b269 -> trunk/c9174a20f75ba007b0f520326ced5c65d8a2b269 2025-07-17T08:34:52.2458876Z * [new tag] trunk/c92f1075aaf3649f6368af2a3df9b5167f941b3f -> trunk/c92f1075aaf3649f6368af2a3df9b5167f941b3f 2025-07-17T08:34:52.2461191Z * [new tag] trunk/c9404faacb7c18af17086be26587b46e141afbcc -> trunk/c9404faacb7c18af17086be26587b46e141afbcc 2025-07-17T08:34:52.2463542Z * [new tag] trunk/c95705dac2da8134c946cdd573944632ef22f230 -> trunk/c95705dac2da8134c946cdd573944632ef22f230 2025-07-17T08:34:52.2465734Z * [new tag] trunk/c95f7fa874a3116f1067f9092456ee7281003614 -> trunk/c95f7fa874a3116f1067f9092456ee7281003614 2025-07-17T08:34:52.2468093Z * [new tag] trunk/c9a5bf09baa1a6776555e901331b1baefe6691bc -> trunk/c9a5bf09baa1a6776555e901331b1baefe6691bc 2025-07-17T08:34:52.2470495Z * [new tag] trunk/c9afcffed0870c211d2e8be422812d6a70aef3e2 -> trunk/c9afcffed0870c211d2e8be422812d6a70aef3e2 2025-07-17T08:34:52.2472841Z * [new tag] trunk/c9e9a0c82316ed59f6ef989a6000f0d5231b9b16 -> trunk/c9e9a0c82316ed59f6ef989a6000f0d5231b9b16 2025-07-17T08:34:52.2478744Z * [new tag] trunk/ca3cabd24ace1899ca4669431f5b0556c5ab9ebf -> trunk/ca3cabd24ace1899ca4669431f5b0556c5ab9ebf 2025-07-17T08:34:52.2480800Z * [new tag] trunk/ca5a40395d8b212dce0a57df01ca056eae55446f -> trunk/ca5a40395d8b212dce0a57df01ca056eae55446f 2025-07-17T08:34:52.2482716Z * [new tag] trunk/cadcb5d36802579441da3dbe3c2eecc74a0eae78 -> trunk/cadcb5d36802579441da3dbe3c2eecc74a0eae78 2025-07-17T08:34:52.2485130Z * [new tag] trunk/cb711c8fa04673d3490306e2b14539ab7dca3c23 -> trunk/cb711c8fa04673d3490306e2b14539ab7dca3c23 2025-07-17T08:34:52.2487620Z * [new tag] trunk/cb853945a77af54d9989ae7ac36d0007c0215b10 -> trunk/cb853945a77af54d9989ae7ac36d0007c0215b10 2025-07-17T08:34:52.2489728Z * [new tag] trunk/cb9b479f4f799eea058de9be3a0e09f977fc82da -> trunk/cb9b479f4f799eea058de9be3a0e09f977fc82da 2025-07-17T08:34:52.2492048Z * [new tag] trunk/cbafba57941c1d5f78640845320ac89157570b44 -> trunk/cbafba57941c1d5f78640845320ac89157570b44 2025-07-17T08:34:52.2494241Z * [new tag] trunk/cbcffce48a3422a5e9a2103888248c82d52d199c -> trunk/cbcffce48a3422a5e9a2103888248c82d52d199c 2025-07-17T08:34:52.2496545Z * [new tag] trunk/cc09d3a5ba2a5ae032078570e6123d523952ed22 -> trunk/cc09d3a5ba2a5ae032078570e6123d523952ed22 2025-07-17T08:34:52.2498763Z * [new tag] trunk/cc0faeb80fff17b3d170aa70041865aafb1790a9 -> trunk/cc0faeb80fff17b3d170aa70041865aafb1790a9 2025-07-17T08:34:52.2500993Z * [new tag] trunk/cc3ea2d84004b6348c77a285ba2639932b87c101 -> trunk/cc3ea2d84004b6348c77a285ba2639932b87c101 2025-07-17T08:34:52.2503361Z * [new tag] trunk/cc410d3761001499611bca6cf835239b86093791 -> trunk/cc410d3761001499611bca6cf835239b86093791 2025-07-17T08:34:52.2505707Z * [new tag] trunk/ccb1f687d63f0215ebcf3c1340c6682edeaccd28 -> trunk/ccb1f687d63f0215ebcf3c1340c6682edeaccd28 2025-07-17T08:34:52.2507873Z * [new tag] trunk/ccb67f39b4d75943cdc531d9032b3ba9235cfd56 -> trunk/ccb67f39b4d75943cdc531d9032b3ba9235cfd56 2025-07-17T08:34:52.2510275Z * [new tag] trunk/ccc6279b4086d55cd1f7e2d699473478610d8a7b -> trunk/ccc6279b4086d55cd1f7e2d699473478610d8a7b 2025-07-17T08:34:52.2512542Z * [new tag] trunk/cce4d213a67b7536dce431fd29c92f9791c9f81d -> trunk/cce4d213a67b7536dce431fd29c92f9791c9f81d 2025-07-17T08:34:52.2515095Z * [new tag] trunk/cd1a924dba145334aacb7c1b7276d6cef7ca41cc -> trunk/cd1a924dba145334aacb7c1b7276d6cef7ca41cc 2025-07-17T08:34:52.2517268Z * [new tag] trunk/cd361fc247a9abdbe9851867e31ac3cefcff299e -> trunk/cd361fc247a9abdbe9851867e31ac3cefcff299e 2025-07-17T08:34:52.2519559Z * [new tag] trunk/cd66ff80307862ef8e75520054ecd19a5eff9f7e -> trunk/cd66ff80307862ef8e75520054ecd19a5eff9f7e 2025-07-17T08:34:52.2521966Z * [new tag] trunk/cd75cf3cabedeac2ae4bbb266c05ca9b64181abe -> trunk/cd75cf3cabedeac2ae4bbb266c05ca9b64181abe 2025-07-17T08:34:52.2524353Z * [new tag] trunk/cd80f9a4c35a63ef415dbf6a361e4520a9cabf7b -> trunk/cd80f9a4c35a63ef415dbf6a361e4520a9cabf7b 2025-07-17T08:34:52.2526612Z * [new tag] trunk/cd82096973985042ed2eaa84e69fbf0f067f9301 -> trunk/cd82096973985042ed2eaa84e69fbf0f067f9301 2025-07-17T08:34:52.2528825Z * [new tag] trunk/cd995bfb2aac8891465809be3ce29543bd524287 -> trunk/cd995bfb2aac8891465809be3ce29543bd524287 2025-07-17T08:34:52.2531125Z * [new tag] trunk/cdb144fcf0aa4300c6de7158cb18ff7503d48492 -> trunk/cdb144fcf0aa4300c6de7158cb18ff7503d48492 2025-07-17T08:34:52.2533365Z * [new tag] trunk/cdfa33a328f2d3fdaeb645fde186194ddc2e66c0 -> trunk/cdfa33a328f2d3fdaeb645fde186194ddc2e66c0 2025-07-17T08:34:52.2535627Z * [new tag] trunk/ce1a07570d95cdd1543b9b2d45b3899c59e6802f -> trunk/ce1a07570d95cdd1543b9b2d45b3899c59e6802f 2025-07-17T08:34:52.2537814Z * [new tag] trunk/ce3406817d50b3357fa644784cc84ff167ce40ce -> trunk/ce3406817d50b3357fa644784cc84ff167ce40ce 2025-07-17T08:34:52.2540228Z * [new tag] trunk/ce44877961af7c8ec618f525853ce7edf3efa4eb -> trunk/ce44877961af7c8ec618f525853ce7edf3efa4eb 2025-07-17T08:34:52.2542635Z * [new tag] trunk/ce6e0523f9911c605860c4f5f278c2dad893dde9 -> trunk/ce6e0523f9911c605860c4f5f278c2dad893dde9 2025-07-17T08:34:52.2545201Z * [new tag] trunk/ce73b0c53f1f215345dcfe3953871a4aaf170a0c -> trunk/ce73b0c53f1f215345dcfe3953871a4aaf170a0c 2025-07-17T08:34:52.2547990Z * [new tag] trunk/ce79056471737557dcc64378985cd2b036e7322c -> trunk/ce79056471737557dcc64378985cd2b036e7322c 2025-07-17T08:34:52.2549417Z * [new tag] trunk/ce8180a61de405c519e00ac48eab8ff765ea71e6 -> trunk/ce8180a61de405c519e00ac48eab8ff765ea71e6 2025-07-17T08:34:52.2551730Z * [new tag] trunk/ce97a5dcfa3cb10c7805ff5cb44abd6a16b4ae8b -> trunk/ce97a5dcfa3cb10c7805ff5cb44abd6a16b4ae8b 2025-07-17T08:34:52.2554122Z * [new tag] trunk/ce9ba071fd29013e72100dd97728d01c860720d9 -> trunk/ce9ba071fd29013e72100dd97728d01c860720d9 2025-07-17T08:34:52.2556505Z * [new tag] trunk/cec264c8c6d74d4fbf5ee31983737b8b5e316dd0 -> trunk/cec264c8c6d74d4fbf5ee31983737b8b5e316dd0 2025-07-17T08:34:52.2563485Z * [new tag] trunk/cec59b76ca606c3e5d34ac0d0f9e0e22b8cfe5bb -> trunk/cec59b76ca606c3e5d34ac0d0f9e0e22b8cfe5bb 2025-07-17T08:34:52.2565637Z * [new tag] trunk/cf0749c92fdfcb8a583ea15f351add88c4d33618 -> trunk/cf0749c92fdfcb8a583ea15f351add88c4d33618 2025-07-17T08:34:52.2567991Z * [new tag] trunk/cf3247b74aaeb956b3c2b31d05e965a0aca5a8b4 -> trunk/cf3247b74aaeb956b3c2b31d05e965a0aca5a8b4 2025-07-17T08:34:52.2570303Z * [new tag] trunk/cf90c9f8d1632777ec5f4b6ccaa14bc5bf259e9c -> trunk/cf90c9f8d1632777ec5f4b6ccaa14bc5bf259e9c 2025-07-17T08:34:52.2572648Z * [new tag] trunk/cf9878d7a20b11d761934ae7732b01644ae137d6 -> trunk/cf9878d7a20b11d761934ae7732b01644ae137d6 2025-07-17T08:34:52.2575072Z * [new tag] trunk/d061a02e6ecf3b62f409578c7d05a564264d1288 -> trunk/d061a02e6ecf3b62f409578c7d05a564264d1288 2025-07-17T08:34:52.2577529Z * [new tag] trunk/d06a406656bf0b4bc9af2d056558dc1630a6e5b9 -> trunk/d06a406656bf0b4bc9af2d056558dc1630a6e5b9 2025-07-17T08:34:52.2579831Z * [new tag] trunk/d083841c0e0d879e460aebb45daac086a240a7dd -> trunk/d083841c0e0d879e460aebb45daac086a240a7dd 2025-07-17T08:34:52.2582589Z * [new tag] trunk/d0a9629435aaceb5acbf31aad70f2109cb8a3ea2 -> trunk/d0a9629435aaceb5acbf31aad70f2109cb8a3ea2 2025-07-17T08:34:52.2584886Z * [new tag] trunk/d0cfa3e5bf662b72068e26e06a1a5c2e640ce8d9 -> trunk/d0cfa3e5bf662b72068e26e06a1a5c2e640ce8d9 2025-07-17T08:34:52.2587162Z * [new tag] trunk/d158e9ea82ce3440e2e9a1bbbecae53a79869404 -> trunk/d158e9ea82ce3440e2e9a1bbbecae53a79869404 2025-07-17T08:34:52.2589467Z * [new tag] trunk/d1947a87074c5db2568038878b1948ea3a33cc23 -> trunk/d1947a87074c5db2568038878b1948ea3a33cc23 2025-07-17T08:34:52.2591743Z * [new tag] trunk/d1b4e0fa9a5feb22fc6de1d36dc4c9dac685caed -> trunk/d1b4e0fa9a5feb22fc6de1d36dc4c9dac685caed 2025-07-17T08:34:52.2594158Z * [new tag] trunk/d1c924c68aaf650439c2193f74cfefc9f16076f1 -> trunk/d1c924c68aaf650439c2193f74cfefc9f16076f1 2025-07-17T08:34:52.2596814Z * [new tag] trunk/d26ca5de058dbcf56ac52bb43e84dd98df2ace97 -> trunk/d26ca5de058dbcf56ac52bb43e84dd98df2ace97 2025-07-17T08:34:52.2599015Z * [new tag] trunk/d27d36136ce35d5d6dc3faa818ba840ba61d4357 -> trunk/d27d36136ce35d5d6dc3faa818ba840ba61d4357 2025-07-17T08:34:52.2601276Z * [new tag] trunk/d283fc79b15844ac470b3429b4dac5661621ae5c -> trunk/d283fc79b15844ac470b3429b4dac5661621ae5c 2025-07-17T08:34:52.2603599Z * [new tag] trunk/d2a2bfcb5867d4920e0a45113861cde6f74857b9 -> trunk/d2a2bfcb5867d4920e0a45113861cde6f74857b9 2025-07-17T08:34:52.2606000Z * [new tag] trunk/d2f06d2b062c9d868a898044bc324547f34c4760 -> trunk/d2f06d2b062c9d868a898044bc324547f34c4760 2025-07-17T08:34:52.2608475Z * [new tag] trunk/d309cd1d502eab8e9cb536876bf5eed88634eb41 -> trunk/d309cd1d502eab8e9cb536876bf5eed88634eb41 2025-07-17T08:34:52.2610555Z * [new tag] trunk/d32deb664a1dfc0e36f0a8d670695046632ebf2a -> trunk/d32deb664a1dfc0e36f0a8d670695046632ebf2a 2025-07-17T08:34:52.2612914Z * [new tag] trunk/d3d64c6db090ee9be051202e6ac6fd4acd5d3e97 -> trunk/d3d64c6db090ee9be051202e6ac6fd4acd5d3e97 2025-07-17T08:34:52.2615254Z * [new tag] trunk/d3d655ad14ee4cd1c135ac57bbf75d5623fc9fa6 -> trunk/d3d655ad14ee4cd1c135ac57bbf75d5623fc9fa6 2025-07-17T08:34:52.2617578Z * [new tag] trunk/d3da03d6fa02aecd01ff41246def793b8fc630bf -> trunk/d3da03d6fa02aecd01ff41246def793b8fc630bf 2025-07-17T08:34:52.2619973Z * [new tag] trunk/d3efd732348f8a455ca23e7fcae043651e41f65e -> trunk/d3efd732348f8a455ca23e7fcae043651e41f65e 2025-07-17T08:34:52.2622276Z * [new tag] trunk/d40aaa42ee8d7283d8de02c68c5625d67acf668a -> trunk/d40aaa42ee8d7283d8de02c68c5625d67acf668a 2025-07-17T08:34:52.2624515Z * [new tag] trunk/d41f62b7a06c51e4a57df4d58e7a2d86d2faa875 -> trunk/d41f62b7a06c51e4a57df4d58e7a2d86d2faa875 2025-07-17T08:34:52.2626747Z * [new tag] trunk/d42c11819f003a848c782d6090f44bbca11c7d60 -> trunk/d42c11819f003a848c782d6090f44bbca11c7d60 2025-07-17T08:34:52.2629072Z * [new tag] trunk/d43c0bdf46d82806cff195a6d2bfa103ab46dc14 -> trunk/d43c0bdf46d82806cff195a6d2bfa103ab46dc14 2025-07-17T08:34:52.2631396Z * [new tag] trunk/d4ad28042993dd09c80775d7fc5756fb4d13fdd4 -> trunk/d4ad28042993dd09c80775d7fc5756fb4d13fdd4 2025-07-17T08:34:52.2633716Z * [new tag] trunk/d4d0ede6bacb4b3b33c0e4aa4cb0e79d34e697ec -> trunk/d4d0ede6bacb4b3b33c0e4aa4cb0e79d34e697ec 2025-07-17T08:34:52.2636156Z * [new tag] trunk/d512584718838876ae14b32bb53cd836a1ceccae -> trunk/d512584718838876ae14b32bb53cd836a1ceccae 2025-07-17T08:34:52.2638322Z * [new tag] trunk/d55dc00f8479699f454e91c779b53e2a16cffc21 -> trunk/d55dc00f8479699f454e91c779b53e2a16cffc21 2025-07-17T08:34:52.2640961Z * [new tag] trunk/d56f11a1f275f2dc037cd5099b70240745990693 -> trunk/d56f11a1f275f2dc037cd5099b70240745990693 2025-07-17T08:34:52.2643205Z * [new tag] trunk/d5781c8d21b3dca35715a093ba52c5698551ad9b -> trunk/d5781c8d21b3dca35715a093ba52c5698551ad9b 2025-07-17T08:34:52.2645484Z * [new tag] trunk/d58ed04d89c34c6930d0f28be351c53db407078f -> trunk/d58ed04d89c34c6930d0f28be351c53db407078f 2025-07-17T08:34:52.2648139Z * [new tag] trunk/d59ed21d0fdcbb34b5c4b121a3b49ae64d8da367 -> trunk/d59ed21d0fdcbb34b5c4b121a3b49ae64d8da367 2025-07-17T08:34:52.2650015Z * [new tag] trunk/d5a89178b05db3124f16c733e9b00f8c5c4ab038 -> trunk/d5a89178b05db3124f16c733e9b00f8c5c4ab038 2025-07-17T08:34:52.2652146Z * [new tag] trunk/d5b4a329607514b30944c7682efcf79cf8950ed3 -> trunk/d5b4a329607514b30944c7682efcf79cf8950ed3 2025-07-17T08:34:52.2654492Z * [new tag] trunk/d5d14ee823e70224a154884a5c2912643d648094 -> trunk/d5d14ee823e70224a154884a5c2912643d648094 2025-07-17T08:34:52.2656877Z * [new tag] trunk/d5e6f4209424b96237ea5468dd47899e1124dfcb -> trunk/d5e6f4209424b96237ea5468dd47899e1124dfcb 2025-07-17T08:34:52.2659483Z * [new tag] trunk/d6237721c074484ea5e72fc05614587886e57fd6 -> trunk/d6237721c074484ea5e72fc05614587886e57fd6 2025-07-17T08:34:52.2661714Z * [new tag] trunk/d632cf2cc9aac8ab0e03d1537982265e42be95e5 -> trunk/d632cf2cc9aac8ab0e03d1537982265e42be95e5 2025-07-17T08:34:52.2664023Z * [new tag] trunk/d66b4bcc3f89ac6b9d2ccfd4cb01b0aa85d4e3fd -> trunk/d66b4bcc3f89ac6b9d2ccfd4cb01b0aa85d4e3fd 2025-07-17T08:34:52.2666129Z * [new tag] trunk/d6ee5144cab20141c8e0cf48856a90a6517753f1 -> trunk/d6ee5144cab20141c8e0cf48856a90a6517753f1 2025-07-17T08:34:52.2668488Z * [new tag] trunk/d75d30eeb610b164e69d0678a2e2b2dea81eec0f -> trunk/d75d30eeb610b164e69d0678a2e2b2dea81eec0f 2025-07-17T08:34:52.2671229Z * [new tag] trunk/d76323d41742cbc05ec6857319b267d2c7ea8fd9 -> trunk/d76323d41742cbc05ec6857319b267d2c7ea8fd9 2025-07-17T08:34:52.2673131Z * [new tag] trunk/d79651571fb082a7a20631059563ecdf9d11b7e6 -> trunk/d79651571fb082a7a20631059563ecdf9d11b7e6 2025-07-17T08:34:52.2675711Z * [new tag] trunk/d797038ea9f3951c7602e8710f4866f73d7e7acd -> trunk/d797038ea9f3951c7602e8710f4866f73d7e7acd 2025-07-17T08:34:52.2677959Z * [new tag] trunk/d7e1b8b11d7430c7633dcad6f6596b5df8fa02f7 -> trunk/d7e1b8b11d7430c7633dcad6f6596b5df8fa02f7 2025-07-17T08:34:52.2680244Z * [new tag] trunk/d7e3c9ce8289f26704b140e68dac2849df750040 -> trunk/d7e3c9ce8289f26704b140e68dac2849df750040 2025-07-17T08:34:52.2682623Z * [new tag] trunk/d7e657da358d6d393e2105f94f1312fb969654a2 -> trunk/d7e657da358d6d393e2105f94f1312fb969654a2 2025-07-17T08:34:52.2684922Z * [new tag] trunk/d83ff89d3bf54ecc95d441411ecbf07831cc0e5a -> trunk/d83ff89d3bf54ecc95d441411ecbf07831cc0e5a 2025-07-17T08:34:52.2687243Z * [new tag] trunk/d846e213553621cb4581f1f36ac0023e528ddb65 -> trunk/d846e213553621cb4581f1f36ac0023e528ddb65 2025-07-17T08:34:52.2689602Z * [new tag] trunk/d84efde3f02bdc4780e5c9ec43f9c515c744e9be -> trunk/d84efde3f02bdc4780e5c9ec43f9c515c744e9be 2025-07-17T08:34:52.2691853Z * [new tag] trunk/d859e65826d511f8379db4a8c4109417e26e0772 -> trunk/d859e65826d511f8379db4a8c4109417e26e0772 2025-07-17T08:34:52.2694234Z * [new tag] trunk/d89f30ad45b9d4bfe5cf5ab441b53e849e55df7b -> trunk/d89f30ad45b9d4bfe5cf5ab441b53e849e55df7b 2025-07-17T08:34:52.2696528Z * [new tag] trunk/d8bb5ac260dd81f75c8add6eadd8ce11e62be1a3 -> trunk/d8bb5ac260dd81f75c8add6eadd8ce11e62be1a3 2025-07-17T08:34:52.2698704Z * [new tag] trunk/d9426a81d2ab54f809a3b32a6ab2e606073fe66f -> trunk/d9426a81d2ab54f809a3b32a6ab2e606073fe66f 2025-07-17T08:34:52.2700844Z * [new tag] trunk/d9577df312d477e8fa5b9d7bc61fb1f2c07b8e48 -> trunk/d9577df312d477e8fa5b9d7bc61fb1f2c07b8e48 2025-07-17T08:34:52.2703217Z * [new tag] trunk/d96dec84151dfdea1aa2bf568deebdc182214f45 -> trunk/d96dec84151dfdea1aa2bf568deebdc182214f45 2025-07-17T08:34:52.2705463Z * [new tag] trunk/d9799a2ee7adb9d15f7a2d47ee17fc39c588af60 -> trunk/d9799a2ee7adb9d15f7a2d47ee17fc39c588af60 2025-07-17T08:34:52.2707590Z * [new tag] trunk/d98fa4a1033079ae19591aff5c3c354617bff482 -> trunk/d98fa4a1033079ae19591aff5c3c354617bff482 2025-07-17T08:34:52.2709869Z * [new tag] trunk/d99cac28160428a8b006890abffaaa0754bd28e1 -> trunk/d99cac28160428a8b006890abffaaa0754bd28e1 2025-07-17T08:34:52.2712266Z * [new tag] trunk/d9b8369f3976e89eef96c903dfede746fb438f63 -> trunk/d9b8369f3976e89eef96c903dfede746fb438f63 2025-07-17T08:34:52.2714533Z * [new tag] trunk/da05b7fb94fa6382c43e165a525a76d8ae62cadd -> trunk/da05b7fb94fa6382c43e165a525a76d8ae62cadd 2025-07-17T08:34:52.2716775Z * [new tag] trunk/da1f337bc43ce2d720aca1349f9c09712e037382 -> trunk/da1f337bc43ce2d720aca1349f9c09712e037382 2025-07-17T08:34:52.2719039Z * [new tag] trunk/da1f8980dfd646f14dc33bf0bc70fcb4d09096b3 -> trunk/da1f8980dfd646f14dc33bf0bc70fcb4d09096b3 2025-07-17T08:34:52.2721320Z * [new tag] trunk/da4e7c77a12746b0806a63af58d7c913ed241f55 -> trunk/da4e7c77a12746b0806a63af58d7c913ed241f55 2025-07-17T08:34:52.2723621Z * [new tag] trunk/da50835bdebd43275e0be4b1362e1bb92b568e72 -> trunk/da50835bdebd43275e0be4b1362e1bb92b568e72 2025-07-17T08:34:52.2725909Z * [new tag] trunk/da910e603a764c1f935dd0409b4c18bd84b1f045 -> trunk/da910e603a764c1f935dd0409b4c18bd84b1f045 2025-07-17T08:34:52.2728253Z * [new tag] trunk/dabb55baff2292ddd7728882b4e2113b8cb9e23b -> trunk/dabb55baff2292ddd7728882b4e2113b8cb9e23b 2025-07-17T08:34:52.2730976Z * [new tag] trunk/db01f1032ff133dc7cc3097b76bfb1b73ff533ce -> trunk/db01f1032ff133dc7cc3097b76bfb1b73ff533ce 2025-07-17T08:34:52.2733111Z * [new tag] trunk/db188503cb705230664c94d8b3a85165a00d632d -> trunk/db188503cb705230664c94d8b3a85165a00d632d 2025-07-17T08:34:52.2735384Z * [new tag] trunk/db491825e05e8bed9c7ed5d9fdd6ce206ab3167c -> trunk/db491825e05e8bed9c7ed5d9fdd6ce206ab3167c 2025-07-17T08:34:52.2737634Z * [new tag] trunk/db5970c1a67968f3b76d204d75789021d4304337 -> trunk/db5970c1a67968f3b76d204d75789021d4304337 2025-07-17T08:34:52.2740195Z * [new tag] trunk/dbec08bc1cc9113c13652128a4ddcfcf2426a8c0 -> trunk/dbec08bc1cc9113c13652128a4ddcfcf2426a8c0 2025-07-17T08:34:52.2742567Z * [new tag] trunk/dbf7d421dabced2335d17c7d7e57c1770f2f12c0 -> trunk/dbf7d421dabced2335d17c7d7e57c1770f2f12c0 2025-07-17T08:34:52.2744906Z * [new tag] trunk/dc5e8f7999cccb51efcf0f5fe197a740a918c73d -> trunk/dc5e8f7999cccb51efcf0f5fe197a740a918c73d 2025-07-17T08:34:52.2747160Z * [new tag] trunk/dcb97cd519e4e4974bc3bd4960512b557216f3e6 -> trunk/dcb97cd519e4e4974bc3bd4960512b557216f3e6 2025-07-17T08:34:52.2749679Z * [new tag] trunk/dd1b6621bc04a37b60ac304d913a3c838b0b1e4d -> trunk/dd1b6621bc04a37b60ac304d913a3c838b0b1e4d 2025-07-17T08:34:52.2751974Z * [new tag] trunk/dd3e7170c20cade9a308900aa4df46ab7656d253 -> trunk/dd3e7170c20cade9a308900aa4df46ab7656d253 2025-07-17T08:34:52.2754336Z * [new tag] trunk/dd41a3907cde3ce19fb69bd79c4055233a1e28d3 -> trunk/dd41a3907cde3ce19fb69bd79c4055233a1e28d3 2025-07-17T08:34:52.2756934Z * [new tag] trunk/dd78d6e7eaecf0aa5317076a721bc719c96b41f8 -> trunk/dd78d6e7eaecf0aa5317076a721bc719c96b41f8 2025-07-17T08:34:52.2759004Z * [new tag] trunk/dd93883231226576220a8ec0fa266c9e22eae82c -> trunk/dd93883231226576220a8ec0fa266c9e22eae82c 2025-07-17T08:34:52.2761286Z * [new tag] trunk/ddf502c988133835a89959bef945bf9c5f06b428 -> trunk/ddf502c988133835a89959bef945bf9c5f06b428 2025-07-17T08:34:52.2763607Z * [new tag] trunk/de1930a4290a27f3f03dd1d3454fb6ef9d12f3ea -> trunk/de1930a4290a27f3f03dd1d3454fb6ef9d12f3ea 2025-07-17T08:34:52.2765962Z * [new tag] trunk/de45c5f673ce261e9a82c54280beeda36cff640e -> trunk/de45c5f673ce261e9a82c54280beeda36cff640e 2025-07-17T08:34:52.2768445Z * [new tag] trunk/dea4864ce08115a0cdc871b84dd3080567d8b5e4 -> trunk/dea4864ce08115a0cdc871b84dd3080567d8b5e4 2025-07-17T08:34:52.2770673Z * [new tag] trunk/df72078fe1339751e702c7511c23b4597d022dcc -> trunk/df72078fe1339751e702c7511c23b4597d022dcc 2025-07-17T08:34:52.2773011Z * [new tag] trunk/dfa2649434f539d7580d38f08890176e73d45158 -> trunk/dfa2649434f539d7580d38f08890176e73d45158 2025-07-17T08:34:52.2775284Z * [new tag] trunk/dfc31b3345d78b0a49d446dcc1957404606a3aa2 -> trunk/dfc31b3345d78b0a49d446dcc1957404606a3aa2 2025-07-17T08:34:52.2777696Z * [new tag] trunk/dfcda613b65499e4b593b5bbb64812d57194858e -> trunk/dfcda613b65499e4b593b5bbb64812d57194858e 2025-07-17T08:34:52.2779962Z * [new tag] trunk/dfdd636cfa06ac70c4aa7ad5ee5e854c46b25a8e -> trunk/dfdd636cfa06ac70c4aa7ad5ee5e854c46b25a8e 2025-07-17T08:34:52.2782292Z * [new tag] trunk/dfef1e44085bb156abc4aff0f34a0b82a4a337b8 -> trunk/dfef1e44085bb156abc4aff0f34a0b82a4a337b8 2025-07-17T08:34:52.2784616Z * [new tag] trunk/e01fde82131c7f0b4c122222694911ee6fab36ca -> trunk/e01fde82131c7f0b4c122222694911ee6fab36ca 2025-07-17T08:34:52.2787038Z * [new tag] trunk/e0447bb5f84dca38e7515d1b1fdea42c647e5acd -> trunk/e0447bb5f84dca38e7515d1b1fdea42c647e5acd 2025-07-17T08:34:52.2789304Z * [new tag] trunk/e071837594f26a3bbaeee11eaa74f3bfd9c998ce -> trunk/e071837594f26a3bbaeee11eaa74f3bfd9c998ce 2025-07-17T08:34:52.2791756Z * [new tag] trunk/e0850123355a919427be3cb02f40c9a9e144fc3c -> trunk/e0850123355a919427be3cb02f40c9a9e144fc3c 2025-07-17T08:34:52.2793842Z * [new tag] trunk/e0ab1b538a646bff0739830d74a8e6ea6bf1eddd -> trunk/e0ab1b538a646bff0739830d74a8e6ea6bf1eddd 2025-07-17T08:34:52.2796344Z * [new tag] trunk/e0ae4ecca87928a96c97fb60cbb36aca0fc182f2 -> trunk/e0ae4ecca87928a96c97fb60cbb36aca0fc182f2 2025-07-17T08:34:52.2798641Z * [new tag] trunk/e0fd48be7d526625cacfd093f3db13506ca9cc17 -> trunk/e0fd48be7d526625cacfd093f3db13506ca9cc17 2025-07-17T08:34:52.2801254Z * [new tag] trunk/e1180c7228ba8c8b16cabf78706d4a67ca189a6b -> trunk/e1180c7228ba8c8b16cabf78706d4a67ca189a6b 2025-07-17T08:34:52.2803720Z * [new tag] trunk/e124a0d88ca2aa04bfaca2dcabf5de6244048e45 -> trunk/e124a0d88ca2aa04bfaca2dcabf5de6244048e45 2025-07-17T08:34:52.2805986Z * [new tag] trunk/e12597090c482ebf1eff7739794667a92cbff657 -> trunk/e12597090c482ebf1eff7739794667a92cbff657 2025-07-17T08:34:52.2808358Z * [new tag] trunk/e15848669f84d3767bfca724a29a6a6dde3308b9 -> trunk/e15848669f84d3767bfca724a29a6a6dde3308b9 2025-07-17T08:34:52.2810618Z * [new tag] trunk/e15ea965a1e84029eb900b93f5776767a39fd91e -> trunk/e15ea965a1e84029eb900b93f5776767a39fd91e 2025-07-17T08:34:52.2812887Z * [new tag] trunk/e15f4248ad2797539f1bf965bd0a3500b8ed15ed -> trunk/e15f4248ad2797539f1bf965bd0a3500b8ed15ed 2025-07-17T08:34:52.2815083Z * [new tag] trunk/e1723098805fcbf3ab7f3d20750fb0665bf65004 -> trunk/e1723098805fcbf3ab7f3d20750fb0665bf65004 2025-07-17T08:34:52.2817488Z * [new tag] trunk/e1a20988f3724317a7ee79c1777d574a8282a122 -> trunk/e1a20988f3724317a7ee79c1777d574a8282a122 2025-07-17T08:34:52.2820112Z * [new tag] trunk/e1aee86646aa6d1b9cb9d34351e43936401c5efc -> trunk/e1aee86646aa6d1b9cb9d34351e43936401c5efc 2025-07-17T08:34:52.2822422Z * [new tag] trunk/e1db10e05aa720aef1989773adcf48f311bcf920 -> trunk/e1db10e05aa720aef1989773adcf48f311bcf920 2025-07-17T08:34:52.2824694Z * [new tag] trunk/e1f28fe17bc55389fc5a31bad588d22e5ecca722 -> trunk/e1f28fe17bc55389fc5a31bad588d22e5ecca722 2025-07-17T08:34:52.2827000Z * [new tag] trunk/e20784f228abca24ab263cf0f69534667365ecba -> trunk/e20784f228abca24ab263cf0f69534667365ecba 2025-07-17T08:34:52.2829278Z * [new tag] trunk/e2351f2dcf7d300460edb9d0c2eb0ea11cc547c5 -> trunk/e2351f2dcf7d300460edb9d0c2eb0ea11cc547c5 2025-07-17T08:34:52.2831681Z * [new tag] trunk/e25ce0f928bfdd545323fcc2f88209233309bbbc -> trunk/e25ce0f928bfdd545323fcc2f88209233309bbbc 2025-07-17T08:34:52.2833926Z * [new tag] trunk/e265b719bd67f7c0a2b9001daef442a70232dcc8 -> trunk/e265b719bd67f7c0a2b9001daef442a70232dcc8 2025-07-17T08:34:52.2836431Z * [new tag] trunk/e28925aa7566da3ffe4f13dd8d59a6767bf97b2e -> trunk/e28925aa7566da3ffe4f13dd8d59a6767bf97b2e 2025-07-17T08:34:52.2838742Z * [new tag] trunk/e290a4c645e00926f0bfa44488f9f7dbd7cb2d0b -> trunk/e290a4c645e00926f0bfa44488f9f7dbd7cb2d0b 2025-07-17T08:34:52.2840997Z * [new tag] trunk/e2c9d8d6414927ce754bbc40b767edf103cf16da -> trunk/e2c9d8d6414927ce754bbc40b767edf103cf16da 2025-07-17T08:34:52.2843313Z * [new tag] trunk/e2f64eedafc6f0dcad9256ac1f38c0359477679a -> trunk/e2f64eedafc6f0dcad9256ac1f38c0359477679a 2025-07-17T08:34:52.2845636Z * [new tag] trunk/e311886e3d57c83a88b97a084dd0b95d6d1537a8 -> trunk/e311886e3d57c83a88b97a084dd0b95d6d1537a8 2025-07-17T08:34:52.2847916Z * [new tag] trunk/e31f20529276356092b5c63c2920d5b17ca9f4ba -> trunk/e31f20529276356092b5c63c2920d5b17ca9f4ba 2025-07-17T08:34:52.2850225Z * [new tag] trunk/e323d46b617c1552cc0f017ad8ec80189eacf187 -> trunk/e323d46b617c1552cc0f017ad8ec80189eacf187 2025-07-17T08:34:52.2852486Z * [new tag] trunk/e3320965b48d87b7257997c35db722302d923ead -> trunk/e3320965b48d87b7257997c35db722302d923ead 2025-07-17T08:34:52.2854984Z * [new tag] trunk/e375d21bb9b0ef6fefe7a8af5a054a17de8c63c9 -> trunk/e375d21bb9b0ef6fefe7a8af5a054a17de8c63c9 2025-07-17T08:34:52.2857138Z * [new tag] trunk/e3977e843de6c9c43be00ee8c67c533debfc0dc9 -> trunk/e3977e843de6c9c43be00ee8c67c533debfc0dc9 2025-07-17T08:34:52.2859394Z * [new tag] trunk/e3afbb03623559fa3fa7ba607c07c86ebed889f0 -> trunk/e3afbb03623559fa3fa7ba607c07c86ebed889f0 2025-07-17T08:34:52.2861975Z * [new tag] trunk/e3b44edfd837199880aed3ec802383617da731f3 -> trunk/e3b44edfd837199880aed3ec802383617da731f3 2025-07-17T08:34:52.2864358Z * [new tag] trunk/e3f2597b4529927b812b443cf8da19b44ed8dea4 -> trunk/e3f2597b4529927b812b443cf8da19b44ed8dea4 2025-07-17T08:34:52.2866653Z * [new tag] trunk/e3f8141c25b7213885f0b7b64f2b0ccc7858be95 -> trunk/e3f8141c25b7213885f0b7b64f2b0ccc7858be95 2025-07-17T08:34:52.2868933Z * [new tag] trunk/e3fe001d9e38153a7797ad25377fdfa34e461113 -> trunk/e3fe001d9e38153a7797ad25377fdfa34e461113 2025-07-17T08:34:52.2871255Z * [new tag] trunk/e40ade5182233f548b25f2732effe3719d16e9ad -> trunk/e40ade5182233f548b25f2732effe3719d16e9ad 2025-07-17T08:34:52.2873578Z * [new tag] trunk/e466dab164d9236bfe5817ec8e4d24c7b9d3e392 -> trunk/e466dab164d9236bfe5817ec8e4d24c7b9d3e392 2025-07-17T08:34:52.2875940Z * [new tag] trunk/e472daa80963aae389089f9dc324b04261e2a5ef -> trunk/e472daa80963aae389089f9dc324b04261e2a5ef 2025-07-17T08:34:52.2878362Z * [new tag] trunk/e4ae60a413b99688d308794baaa1c685b5c19025 -> trunk/e4ae60a413b99688d308794baaa1c685b5c19025 2025-07-17T08:34:52.2880700Z * [new tag] trunk/e4c17d5e1ccd0e730caef484af291243bc1d9cde -> trunk/e4c17d5e1ccd0e730caef484af291243bc1d9cde 2025-07-17T08:34:52.2883156Z * [new tag] trunk/e4c9f6d9a286bd367ce3a157428f588e267d2630 -> trunk/e4c9f6d9a286bd367ce3a157428f588e267d2630 2025-07-17T08:34:52.2885435Z * [new tag] trunk/e517066f41342f0f01043121dcb8ac6b235ee40c -> trunk/e517066f41342f0f01043121dcb8ac6b235ee40c 2025-07-17T08:34:52.2887744Z * [new tag] trunk/e53ddaf1f6ef3f1d22075ad33a5922fb8b277fe0 -> trunk/e53ddaf1f6ef3f1d22075ad33a5922fb8b277fe0 2025-07-17T08:34:52.2890019Z * [new tag] trunk/e581f015eebf483eb6f19447caf2d2452a602c2f -> trunk/e581f015eebf483eb6f19447caf2d2452a602c2f 2025-07-17T08:34:52.2892370Z * [new tag] trunk/e583b888194c8c74ebbd332c09c394acebbbbcff -> trunk/e583b888194c8c74ebbd332c09c394acebbbbcff 2025-07-17T08:34:52.2894605Z * [new tag] trunk/e5a0b73ce9e7bc489150932b60f673155d6ddc30 -> trunk/e5a0b73ce9e7bc489150932b60f673155d6ddc30 2025-07-17T08:34:52.2896855Z * [new tag] trunk/e5a11971919009b5e5bd050b3ed52e324f1ce894 -> trunk/e5a11971919009b5e5bd050b3ed52e324f1ce894 2025-07-17T08:34:52.2899313Z * [new tag] trunk/e5ea24fb27e0ef5dbb0c4039d88c0f4faa221fef -> trunk/e5ea24fb27e0ef5dbb0c4039d88c0f4faa221fef 2025-07-17T08:34:52.2901623Z * [new tag] trunk/e5ed267f830ce9fc6579f928082e563fbed37e3f -> trunk/e5ed267f830ce9fc6579f928082e563fbed37e3f 2025-07-17T08:34:52.2903886Z * [new tag] trunk/e5edd013ab418b8b3609cb3cb1df3804b69d8eef -> trunk/e5edd013ab418b8b3609cb3cb1df3804b69d8eef 2025-07-17T08:34:52.2906328Z * [new tag] trunk/e5f6ffd8109bee99416d9b127f961ab0e253a1c2 -> trunk/e5f6ffd8109bee99416d9b127f961ab0e253a1c2 2025-07-17T08:34:52.2908616Z * [new tag] trunk/e600e044a770d29d1fe5d9638b274a7d4f22f969 -> trunk/e600e044a770d29d1fe5d9638b274a7d4f22f969 2025-07-17T08:34:52.2910897Z * [new tag] trunk/e6252f62efa5e6e4ab0d56967d7461f1e8a7eb7c -> trunk/e6252f62efa5e6e4ab0d56967d7461f1e8a7eb7c 2025-07-17T08:34:52.2913161Z * [new tag] trunk/e694280d1215caf70f41575f2611bfa26c69ebdb -> trunk/e694280d1215caf70f41575f2611bfa26c69ebdb 2025-07-17T08:34:52.2915711Z * [new tag] trunk/e6d71f37896e99ced7daf9ee43a30c0ec03d60f2 -> trunk/e6d71f37896e99ced7daf9ee43a30c0ec03d60f2 2025-07-17T08:34:52.2918104Z * [new tag] trunk/e6d8ed02cbeabe961d1a5303973a6d89ee851736 -> trunk/e6d8ed02cbeabe961d1a5303973a6d89ee851736 2025-07-17T08:34:52.2920394Z * [new tag] trunk/e6ed4074e8411000eea2f0de0c0829d909e4701f -> trunk/e6ed4074e8411000eea2f0de0c0829d909e4701f 2025-07-17T08:34:52.2922751Z * [new tag] trunk/e71bb021b9553ddc2db6cb8ea7bf8643552f09fc -> trunk/e71bb021b9553ddc2db6cb8ea7bf8643552f09fc 2025-07-17T08:34:52.2925003Z * [new tag] trunk/e7698ff5cf40729d11df6c32c6df0a163e5d0ce0 -> trunk/e7698ff5cf40729d11df6c32c6df0a163e5d0ce0 2025-07-17T08:34:52.2927366Z * [new tag] trunk/e78f2ac92b709a060aa323d6e527ec2ecc36fb93 -> trunk/e78f2ac92b709a060aa323d6e527ec2ecc36fb93 2025-07-17T08:34:52.2929787Z * [new tag] trunk/e7a66166ce5294a52bf3966aef49562f94343475 -> trunk/e7a66166ce5294a52bf3966aef49562f94343475 2025-07-17T08:34:52.2932091Z * [new tag] trunk/e7da21806fbc96d47ac03ea8f6507deb0ab48481 -> trunk/e7da21806fbc96d47ac03ea8f6507deb0ab48481 2025-07-17T08:34:52.2934437Z * [new tag] trunk/e8217ad8becd2b297682c685a9179997cb0a98cc -> trunk/e8217ad8becd2b297682c685a9179997cb0a98cc 2025-07-17T08:34:52.2936843Z * [new tag] trunk/e895e9689c625cbcd8f46880115e0d093713fa37 -> trunk/e895e9689c625cbcd8f46880115e0d093713fa37 2025-07-17T08:34:52.2939152Z * [new tag] trunk/e8b3dfa7c0b6d74fc84714dbb79cb6a8bad43ae7 -> trunk/e8b3dfa7c0b6d74fc84714dbb79cb6a8bad43ae7 2025-07-17T08:34:52.2941388Z * [new tag] trunk/e8bfce9a43960f417a36636a0e7f6a58b7dc56d2 -> trunk/e8bfce9a43960f417a36636a0e7f6a58b7dc56d2 2025-07-17T08:34:52.2943714Z * [new tag] trunk/e8cf5ff5641c1ce4c3046a85df55b15eaaea38b2 -> trunk/e8cf5ff5641c1ce4c3046a85df55b15eaaea38b2 2025-07-17T08:34:52.2946069Z * [new tag] trunk/e8d29c45e02402bcb4ae7726abd415bebd441852 -> trunk/e8d29c45e02402bcb4ae7726abd415bebd441852 2025-07-17T08:34:52.2948306Z * [new tag] trunk/e90148c91d3b8a759db558374c8539e47338926b -> trunk/e90148c91d3b8a759db558374c8539e47338926b 2025-07-17T08:34:52.2950756Z * [new tag] trunk/e92e3eaf4eb815ea28db9a5af9d9ee48c3f7be3f -> trunk/e92e3eaf4eb815ea28db9a5af9d9ee48c3f7be3f 2025-07-17T08:34:52.2953087Z * [new tag] trunk/e9367a7a4288e626f01fada3912d68756f1ca6d3 -> trunk/e9367a7a4288e626f01fada3912d68756f1ca6d3 2025-07-17T08:34:52.2955977Z * [new tag] trunk/e959dd017d7dcd7c6351c9ef303f3667aebe11f5 -> trunk/e959dd017d7dcd7c6351c9ef303f3667aebe11f5 2025-07-17T08:34:52.2958220Z * [new tag] trunk/e95e8eed0a96f024dc84012494f465b34eae9d22 -> trunk/e95e8eed0a96f024dc84012494f465b34eae9d22 2025-07-17T08:34:52.2960508Z * [new tag] trunk/e96f530af5bf8ceaeaa87ec7f4651490a25bce07 -> trunk/e96f530af5bf8ceaeaa87ec7f4651490a25bce07 2025-07-17T08:34:52.2962808Z * [new tag] trunk/e98dd95446e009ace1722498effbf32250d623e4 -> trunk/e98dd95446e009ace1722498effbf32250d623e4 2025-07-17T08:34:52.2965145Z * [new tag] trunk/e99a2a2dba90c9b6276219104bc99aff2f67e2f7 -> trunk/e99a2a2dba90c9b6276219104bc99aff2f67e2f7 2025-07-17T08:34:52.2967430Z * [new tag] trunk/e99cc126a4f4bd5233ff363f27d14a0fff4b4392 -> trunk/e99cc126a4f4bd5233ff363f27d14a0fff4b4392 2025-07-17T08:34:52.2969683Z * [new tag] trunk/e9fdaf8701b599fd943bb899639b5e8a4966b3c3 -> trunk/e9fdaf8701b599fd943bb899639b5e8a4966b3c3 2025-07-17T08:34:52.2971967Z * [new tag] trunk/ea23eb4b98ae8af40f0c38538a9e63f1b0dc40e2 -> trunk/ea23eb4b98ae8af40f0c38538a9e63f1b0dc40e2 2025-07-17T08:34:52.2974265Z * [new tag] trunk/ea37f720995616d8570045c63486198a021f8285 -> trunk/ea37f720995616d8570045c63486198a021f8285 2025-07-17T08:34:52.2976501Z * [new tag] trunk/ea74fdd24aa7d98433231f4a3d75cfd241d8720e -> trunk/ea74fdd24aa7d98433231f4a3d75cfd241d8720e 2025-07-17T08:34:52.2978934Z * [new tag] trunk/eab45643f22e58ee12d95d8b0162d51ca0a50801 -> trunk/eab45643f22e58ee12d95d8b0162d51ca0a50801 2025-07-17T08:34:52.2981142Z * [new tag] trunk/eabf7cd3c552e6c17321a63d7e5a19da92909910 -> trunk/eabf7cd3c552e6c17321a63d7e5a19da92909910 2025-07-17T08:34:52.2983437Z * [new tag] trunk/eaceb243df6fc17a37a2c1d9f8775d1561a0f67c -> trunk/eaceb243df6fc17a37a2c1d9f8775d1561a0f67c 2025-07-17T08:34:52.2985732Z * [new tag] trunk/eaf32fffb7e83b5331951b254ad213ceb2ae72f8 -> trunk/eaf32fffb7e83b5331951b254ad213ceb2ae72f8 2025-07-17T08:34:52.2988059Z * [new tag] trunk/eaf704914eb63efffafeb402633f76f2cd521ef4 -> trunk/eaf704914eb63efffafeb402633f76f2cd521ef4 2025-07-17T08:34:52.2990366Z * [new tag] trunk/eb152ab1dd9eafc777b6642bb141ab1b8376359a -> trunk/eb152ab1dd9eafc777b6642bb141ab1b8376359a 2025-07-17T08:34:52.2992751Z * [new tag] trunk/eb2af14f8e75c22c0a2fcf407005d2e8173d0df5 -> trunk/eb2af14f8e75c22c0a2fcf407005d2e8173d0df5 2025-07-17T08:34:52.2995108Z * [new tag] trunk/eb331b59fedb7f70229ca491029f2f7f27bbff54 -> trunk/eb331b59fedb7f70229ca491029f2f7f27bbff54 2025-07-17T08:34:52.2997399Z * [new tag] trunk/eb4cf59ecdf69dd53805a37525a7f750c8f1e0dd -> trunk/eb4cf59ecdf69dd53805a37525a7f750c8f1e0dd 2025-07-17T08:34:52.2999715Z * [new tag] trunk/eb9efb37c8f315f1d30e86d5797490c6a8666889 -> trunk/eb9efb37c8f315f1d30e86d5797490c6a8666889 2025-07-17T08:34:52.3001994Z * [new tag] trunk/eba5fc91ac924a7a05569c33b0f90f0faf22696d -> trunk/eba5fc91ac924a7a05569c33b0f90f0faf22696d 2025-07-17T08:34:52.3004260Z * [new tag] trunk/ebab2799423987a16975ab07ee49dd944e5315e6 -> trunk/ebab2799423987a16975ab07ee49dd944e5315e6 2025-07-17T08:34:52.3006542Z * [new tag] trunk/ebf83b8b7772632c0558db9a88281ee10ff2df38 -> trunk/ebf83b8b7772632c0558db9a88281ee10ff2df38 2025-07-17T08:34:52.3008798Z * [new tag] trunk/ec0276103fc023b93c6eb17639d4605656ea4f10 -> trunk/ec0276103fc023b93c6eb17639d4605656ea4f10 2025-07-17T08:34:52.3011194Z * [new tag] trunk/ec08eb8ba22e66b113e4f2aba1f6afb738f9d861 -> trunk/ec08eb8ba22e66b113e4f2aba1f6afb738f9d861 2025-07-17T08:34:52.3013313Z * [new tag] trunk/ec816d73b4c8f74a14e275891939de6cfd6891b6 -> trunk/ec816d73b4c8f74a14e275891939de6cfd6891b6 2025-07-17T08:34:52.3015654Z * [new tag] trunk/ecd73c58eeaf7e919316f9b9596f8c677af96c66 -> trunk/ecd73c58eeaf7e919316f9b9596f8c677af96c66 2025-07-17T08:34:52.3018053Z * [new tag] trunk/ed03492238c9ef76f5c1df6cfa54da83a2a384e1 -> trunk/ed03492238c9ef76f5c1df6cfa54da83a2a384e1 2025-07-17T08:34:52.3020376Z * [new tag] trunk/ed051c308464904e07d33c39524b3931dc8c947a -> trunk/ed051c308464904e07d33c39524b3931dc8c947a 2025-07-17T08:34:52.3022742Z * [new tag] trunk/ed508cc0182b923789555d4e1ec1f8b23707b7db -> trunk/ed508cc0182b923789555d4e1ec1f8b23707b7db 2025-07-17T08:34:52.3025038Z * [new tag] trunk/ed5d6d2a200a0f03b164053322341bb10acc051d -> trunk/ed5d6d2a200a0f03b164053322341bb10acc051d 2025-07-17T08:34:52.3027312Z * [new tag] trunk/ed661a5f11de116539f70882f4f59a61d5e3e209 -> trunk/ed661a5f11de116539f70882f4f59a61d5e3e209 2025-07-17T08:34:52.3029725Z * [new tag] trunk/ed6ae20cf0e31d49d54177251293267205e24021 -> trunk/ed6ae20cf0e31d49d54177251293267205e24021 2025-07-17T08:34:52.3032085Z * [new tag] trunk/ed6df0e324233fda094e3c1cb4de4009d1ba2413 -> trunk/ed6df0e324233fda094e3c1cb4de4009d1ba2413 2025-07-17T08:34:52.3034638Z * [new tag] trunk/ed911747c2a56cf84d648d677403d5bf6b2ebd0b -> trunk/ed911747c2a56cf84d648d677403d5bf6b2ebd0b 2025-07-17T08:34:52.3036976Z * [new tag] trunk/eda0a9cc90b9a63127a49d617329f98b6404e90d -> trunk/eda0a9cc90b9a63127a49d617329f98b6404e90d 2025-07-17T08:34:52.3039656Z * [new tag] trunk/edb92e16ba9ad12bd51c4f73389556e163e9cc08 -> trunk/edb92e16ba9ad12bd51c4f73389556e163e9cc08 2025-07-17T08:34:52.3041759Z * [new tag] trunk/edd45f3a020f892c17672cc2d08f64cb960006ad -> trunk/edd45f3a020f892c17672cc2d08f64cb960006ad 2025-07-17T08:34:52.3044107Z * [new tag] trunk/edd9c09e73aaa8d1a1667fa7ea5a337aa31fa6ec -> trunk/edd9c09e73aaa8d1a1667fa7ea5a337aa31fa6ec 2025-07-17T08:34:52.3046439Z * [new tag] trunk/eddddea9087a8840ce5ae1bbb3c6f59556d7c0f0 -> trunk/eddddea9087a8840ce5ae1bbb3c6f59556d7c0f0 2025-07-17T08:34:52.3048856Z * [new tag] trunk/ede6ead8cd8e925cb093f2b3016342e645bd728d -> trunk/ede6ead8cd8e925cb093f2b3016342e645bd728d 2025-07-17T08:34:52.3050977Z * [new tag] trunk/edf7bb4f514220f96ddfa646ae6e9e930a305ec1 -> trunk/edf7bb4f514220f96ddfa646ae6e9e930a305ec1 2025-07-17T08:34:52.3053117Z * [new tag] trunk/ee0992871c99fc6a1e19eb839ab65391a168d2f8 -> trunk/ee0992871c99fc6a1e19eb839ab65391a168d2f8 2025-07-17T08:34:52.3055545Z * [new tag] trunk/ee4d343499c80be16a58d5ac604da6e2130cd94d -> trunk/ee4d343499c80be16a58d5ac604da6e2130cd94d 2025-07-17T08:34:52.3057865Z * [new tag] trunk/ee56e9f8a8202bfa02c2d9ae3cfa07d4a41ab567 -> trunk/ee56e9f8a8202bfa02c2d9ae3cfa07d4a41ab567 2025-07-17T08:34:52.3060200Z * [new tag] trunk/ee5c2908cbfcbda6e8a5737d9939d36ba417d627 -> trunk/ee5c2908cbfcbda6e8a5737d9939d36ba417d627 2025-07-17T08:34:52.3062470Z * [new tag] trunk/ee9ac36c2316f430c4de2a0982bde3cb8087cbe7 -> trunk/ee9ac36c2316f430c4de2a0982bde3cb8087cbe7 2025-07-17T08:34:52.3064762Z * [new tag] trunk/eea3bcb3d146f16a8f8c91a9e2a8f1b76c5917f4 -> trunk/eea3bcb3d146f16a8f8c91a9e2a8f1b76c5917f4 2025-07-17T08:34:52.3067306Z * [new tag] trunk/eeaefa133694811aac1fd215bcae2e7f1e098d1d -> trunk/eeaefa133694811aac1fd215bcae2e7f1e098d1d 2025-07-17T08:34:52.3069663Z * [new tag] trunk/eecaa0bbc6a39c89366f9ec97797479c27f6d760 -> trunk/eecaa0bbc6a39c89366f9ec97797479c27f6d760 2025-07-17T08:34:52.3072011Z * [new tag] trunk/eef253d9f6fc8b9774d199889a817f50e89572c2 -> trunk/eef253d9f6fc8b9774d199889a817f50e89572c2 2025-07-17T08:34:52.3074564Z * [new tag] trunk/ef4cca2d79eba61441da46906b30f8f6165cc455 -> trunk/ef4cca2d79eba61441da46906b30f8f6165cc455 2025-07-17T08:34:52.3076852Z * [new tag] trunk/ef6d2cee7a93c1e8f52a7595b6f630d0fcd83692 -> trunk/ef6d2cee7a93c1e8f52a7595b6f630d0fcd83692 2025-07-17T08:34:52.3079148Z * [new tag] trunk/ef6dfa06a9ff84bf04b0d83cfab4cb396a508a0e -> trunk/ef6dfa06a9ff84bf04b0d83cfab4cb396a508a0e 2025-07-17T08:34:52.3081399Z * [new tag] trunk/ef97bd47131423e0819b293dc227b62d0c376023 -> trunk/ef97bd47131423e0819b293dc227b62d0c376023 2025-07-17T08:34:52.3083739Z * [new tag] trunk/efbf07e7ea46c33c936cac882eb1c8b782d1d070 -> trunk/efbf07e7ea46c33c936cac882eb1c8b782d1d070 2025-07-17T08:34:52.3086147Z * [new tag] trunk/effe376db07c7071fa4d704dd0787a394a5eda50 -> trunk/effe376db07c7071fa4d704dd0787a394a5eda50 2025-07-17T08:34:52.3088451Z * [new tag] trunk/f04fd4dc4eaf16557f9212240bdaa8377d51bad1 -> trunk/f04fd4dc4eaf16557f9212240bdaa8377d51bad1 2025-07-17T08:34:52.3090699Z * [new tag] trunk/f096820d0f845f36529fe774cf761d5fc4ad02a8 -> trunk/f096820d0f845f36529fe774cf761d5fc4ad02a8 2025-07-17T08:34:52.3092962Z * [new tag] trunk/f0b388665efd20c1ec35fc48afb115a0f15116ce -> trunk/f0b388665efd20c1ec35fc48afb115a0f15116ce 2025-07-17T08:34:52.3095275Z * [new tag] trunk/f0bee87eea03f6ded538590e2a073cc62ecd2818 -> trunk/f0bee87eea03f6ded538590e2a073cc62ecd2818 2025-07-17T08:34:52.3097675Z * [new tag] trunk/f1331f3f1b43d1848341a0f0da66a13cb05570d0 -> trunk/f1331f3f1b43d1848341a0f0da66a13cb05570d0 2025-07-17T08:34:52.3100007Z * [new tag] trunk/f140fac8dcfa6ec20d594bc4637417a83bc90036 -> trunk/f140fac8dcfa6ec20d594bc4637417a83bc90036 2025-07-17T08:34:52.3102430Z * [new tag] trunk/f151b201236f959e3874b73dde9bfae5e10dae78 -> trunk/f151b201236f959e3874b73dde9bfae5e10dae78 2025-07-17T08:34:52.3104739Z * [new tag] trunk/f154f9b3040369a7979d5de7acb6fe21433eda83 -> trunk/f154f9b3040369a7979d5de7acb6fe21433eda83 2025-07-17T08:34:52.3107373Z * [new tag] trunk/f16053f0c9a09fa337fbf85aaf64f88712b8dcdb -> trunk/f16053f0c9a09fa337fbf85aaf64f88712b8dcdb 2025-07-17T08:34:52.3109672Z * [new tag] trunk/f179b7198522e6d93bd103efba1a1ebd5a2cf891 -> trunk/f179b7198522e6d93bd103efba1a1ebd5a2cf891 2025-07-17T08:34:52.3111976Z * [new tag] trunk/f17f6581252ff07e75f33fdce0b4248acdb5da0b -> trunk/f17f6581252ff07e75f33fdce0b4248acdb5da0b 2025-07-17T08:34:52.3114308Z * [new tag] trunk/f1f49e56b079bfc73b1b2fb061d150ec787ee716 -> trunk/f1f49e56b079bfc73b1b2fb061d150ec787ee716 2025-07-17T08:34:52.3116862Z * [new tag] trunk/f2b44424a1f3481ddc066e8dd62c0cafa1298e92 -> trunk/f2b44424a1f3481ddc066e8dd62c0cafa1298e92 2025-07-17T08:34:52.3119145Z * [new tag] trunk/f2d70898c6fad6c6e867d9bf60eced6aacfa9782 -> trunk/f2d70898c6fad6c6e867d9bf60eced6aacfa9782 2025-07-17T08:34:52.3121518Z * [new tag] trunk/f2e712ca14dbbb8435427fd536b0dd63994a9265 -> trunk/f2e712ca14dbbb8435427fd536b0dd63994a9265 2025-07-17T08:34:52.3123950Z * [new tag] trunk/f2ecf6145fde55baa8a91e27b6b3489172f0e639 -> trunk/f2ecf6145fde55baa8a91e27b6b3489172f0e639 2025-07-17T08:34:52.3126242Z * [new tag] trunk/f34335bf3373301ba3af3ff8268cd0cd1715bd2e -> trunk/f34335bf3373301ba3af3ff8268cd0cd1715bd2e 2025-07-17T08:34:52.3128528Z * [new tag] trunk/f34ab1628b6798cc62f55260dfad872f5df8fd94 -> trunk/f34ab1628b6798cc62f55260dfad872f5df8fd94 2025-07-17T08:34:52.3131034Z * [new tag] trunk/f3e6c8e834a49f738e329b802b4b10912e33915f -> trunk/f3e6c8e834a49f738e329b802b4b10912e33915f 2025-07-17T08:34:52.3133330Z * [new tag] trunk/f3ec16c26a24bcbdcb96c692e321be1a4464067c -> trunk/f3ec16c26a24bcbdcb96c692e321be1a4464067c 2025-07-17T08:34:52.3135575Z * [new tag] trunk/f402eed4d9515dee5bb9b11a2b157de57a5988c1 -> trunk/f402eed4d9515dee5bb9b11a2b157de57a5988c1 2025-07-17T08:34:52.3137840Z * [new tag] trunk/f40efde2a474ffd77c021686da2f14864d6aeb6d -> trunk/f40efde2a474ffd77c021686da2f14864d6aeb6d 2025-07-17T08:34:52.3140332Z * [new tag] trunk/f41d017aa6ca1bd121cee6e428875b7fd31a7ad0 -> trunk/f41d017aa6ca1bd121cee6e428875b7fd31a7ad0 2025-07-17T08:34:52.3142724Z * [new tag] trunk/f4376cac54978b9d7c859c1e3495a57d72ab9601 -> trunk/f4376cac54978b9d7c859c1e3495a57d72ab9601 2025-07-17T08:34:52.3145009Z * [new tag] trunk/f4406689b8805ed23dfef34c6320bb99ea6d3767 -> trunk/f4406689b8805ed23dfef34c6320bb99ea6d3767 2025-07-17T08:34:52.3147313Z * [new tag] trunk/f44a9eee4778d2a2aafe3a0beeea7a37ab8d380e -> trunk/f44a9eee4778d2a2aafe3a0beeea7a37ab8d380e 2025-07-17T08:34:52.3149780Z * [new tag] trunk/f45f483884ef524f2d8260dcf2b2ba2c203eab2f -> trunk/f45f483884ef524f2d8260dcf2b2ba2c203eab2f 2025-07-17T08:34:52.3152092Z * [new tag] trunk/f45f6e86b9ef3d9c94b87c2240dbe9a07bb9fd3d -> trunk/f45f6e86b9ef3d9c94b87c2240dbe9a07bb9fd3d 2025-07-17T08:34:52.3154514Z * [new tag] trunk/f48a1576606186cf2104c82be367b617912322de -> trunk/f48a1576606186cf2104c82be367b617912322de 2025-07-17T08:34:52.3156904Z * [new tag] trunk/f4d60a68dd2fb9fda92af83df016f4cfe3af00ba -> trunk/f4d60a68dd2fb9fda92af83df016f4cfe3af00ba 2025-07-17T08:34:52.3159115Z * [new tag] trunk/f56bfb3030acff03e21d568089f3af9e09ec7cb2 -> trunk/f56bfb3030acff03e21d568089f3af9e09ec7cb2 2025-07-17T08:34:52.3161590Z * [new tag] trunk/f58a680d09e13658a52c6ba05c63c15759846bcc -> trunk/f58a680d09e13658a52c6ba05c63c15759846bcc 2025-07-17T08:34:52.3164097Z * [new tag] trunk/f59c76b5494f83abbb2ca169e13201439396c1aa -> trunk/f59c76b5494f83abbb2ca169e13201439396c1aa 2025-07-17T08:34:52.3166222Z * [new tag] trunk/f5bbaa22536437e9f689db8ee7a5ca472fdf4bd0 -> trunk/f5bbaa22536437e9f689db8ee7a5ca472fdf4bd0 2025-07-17T08:34:52.3168452Z * [new tag] trunk/f5e1b24945cf7852a1425923ca543e1f83be14b1 -> trunk/f5e1b24945cf7852a1425923ca543e1f83be14b1 2025-07-17T08:34:52.3170723Z * [new tag] trunk/f5e6e52f2504d0d2c553fd292c89d999b2033616 -> trunk/f5e6e52f2504d0d2c553fd292c89d999b2033616 2025-07-17T08:34:52.3172944Z * [new tag] trunk/f5eb42e4c0405141134e9c52919ebb36ec77ca5f -> trunk/f5eb42e4c0405141134e9c52919ebb36ec77ca5f 2025-07-17T08:34:52.3175294Z * [new tag] trunk/f5f4beaf562911b5cf1a9e8b9eae5e5946681f3c -> trunk/f5f4beaf562911b5cf1a9e8b9eae5e5946681f3c 2025-07-17T08:34:52.3177593Z * [new tag] trunk/f6d138807f138868de0397936e2bee482c1fb987 -> trunk/f6d138807f138868de0397936e2bee482c1fb987 2025-07-17T08:34:52.3180037Z * [new tag] trunk/f6e18bc1054624bb148632a85c10371d6cc62492 -> trunk/f6e18bc1054624bb148632a85c10371d6cc62492 2025-07-17T08:34:52.3182321Z * [new tag] trunk/f70c80105ebc2a118af848c80a18d6efff820f72 -> trunk/f70c80105ebc2a118af848c80a18d6efff820f72 2025-07-17T08:34:52.3184604Z * [new tag] trunk/f7127b9b940a98596599acda1f89fc5153635a5d -> trunk/f7127b9b940a98596599acda1f89fc5153635a5d 2025-07-17T08:34:52.3186842Z * [new tag] trunk/f7130c097efa826313df44f0dcfa7d4d2e4253ec -> trunk/f7130c097efa826313df44f0dcfa7d4d2e4253ec 2025-07-17T08:34:52.3188974Z * [new tag] trunk/f742b32a2ff6326f67512d2b426cdc2da8413a75 -> trunk/f742b32a2ff6326f67512d2b426cdc2da8413a75 2025-07-17T08:34:52.3191123Z * [new tag] trunk/f79689bd3d145e76746e4cf81e99c40df8272a72 -> trunk/f79689bd3d145e76746e4cf81e99c40df8272a72 2025-07-17T08:34:52.3193505Z * [new tag] trunk/f7a5ad6c2987ee5a83aa5d868cee3b8067d3de94 -> trunk/f7a5ad6c2987ee5a83aa5d868cee3b8067d3de94 2025-07-17T08:34:52.3197553Z * [new tag] trunk/f80a61adf57badf999387efe47b8b2ff3388bd96 -> trunk/f80a61adf57badf999387efe47b8b2ff3388bd96 2025-07-17T08:34:52.3199941Z * [new tag] trunk/f810480dbefabbff6cf0852c9f610f84dd440b8d -> trunk/f810480dbefabbff6cf0852c9f610f84dd440b8d 2025-07-17T08:34:52.3202177Z * [new tag] trunk/f810e98143b7b1fe3436d5315101b22aa8438775 -> trunk/f810e98143b7b1fe3436d5315101b22aa8438775 2025-07-17T08:34:52.3204455Z * [new tag] trunk/f8293116f55a9ad399e4938f0fe3e34c6faa47c4 -> trunk/f8293116f55a9ad399e4938f0fe3e34c6faa47c4 2025-07-17T08:34:52.3207000Z * [new tag] trunk/f85954e043a5b6affb589012e8cc5eff8fdb5358 -> trunk/f85954e043a5b6affb589012e8cc5eff8fdb5358 2025-07-17T08:34:52.3209302Z * [new tag] trunk/f860992db5601a78f73eefc0a56af1d7158d6953 -> trunk/f860992db5601a78f73eefc0a56af1d7158d6953 2025-07-17T08:34:52.3211521Z * [new tag] trunk/f87d1179391d66854e3c6ca20717803cfa22f878 -> trunk/f87d1179391d66854e3c6ca20717803cfa22f878 2025-07-17T08:34:52.3213822Z * [new tag] trunk/f88d7a7a34d5a54d58f0a7edc01ff69f46221b6c -> trunk/f88d7a7a34d5a54d58f0a7edc01ff69f46221b6c 2025-07-17T08:34:52.3216252Z * [new tag] trunk/f8baec8984ed90b526fdc03eec19d8039e7d373b -> trunk/f8baec8984ed90b526fdc03eec19d8039e7d373b 2025-07-17T08:34:52.3218689Z * [new tag] trunk/f8c0a4bd28087b02958b92d7b4f41ebc607292b7 -> trunk/f8c0a4bd28087b02958b92d7b4f41ebc607292b7 2025-07-17T08:34:52.3220999Z * [new tag] trunk/f8cc4c0af802269fbe16a418386a94b3b6547c74 -> trunk/f8cc4c0af802269fbe16a418386a94b3b6547c74 2025-07-17T08:34:52.3223251Z * [new tag] trunk/f97f03c7efcf2b7a45384b9094eb6be4cb419546 -> trunk/f97f03c7efcf2b7a45384b9094eb6be4cb419546 2025-07-17T08:34:52.3225653Z * [new tag] trunk/fa0ea57f5e083daab1eb9cda39ab53a5159b077d -> trunk/fa0ea57f5e083daab1eb9cda39ab53a5159b077d 2025-07-17T08:34:52.3227666Z * [new tag] trunk/fa1c20ae9285f7994a73d2d06025065f96b67a57 -> trunk/fa1c20ae9285f7994a73d2d06025065f96b67a57 2025-07-17T08:34:52.3230048Z * [new tag] trunk/fa3c38c7ae17b8d8fccd0958831f9f1ced9e46b3 -> trunk/fa3c38c7ae17b8d8fccd0958831f9f1ced9e46b3 2025-07-17T08:34:52.3232270Z * [new tag] trunk/fa4f07b5b80bdcf99a1c7452de41939d2ab5886f -> trunk/fa4f07b5b80bdcf99a1c7452de41939d2ab5886f 2025-07-17T08:34:52.3234609Z * [new tag] trunk/fa705f791249c5eee912096f1d7ac7b2d5b60e8c -> trunk/fa705f791249c5eee912096f1d7ac7b2d5b60e8c 2025-07-17T08:34:52.3237080Z * [new tag] trunk/fab53dfdf1d89cecd5e82b12cced9b6dd217e87c -> trunk/fab53dfdf1d89cecd5e82b12cced9b6dd217e87c 2025-07-17T08:34:52.3239347Z * [new tag] trunk/fab85fc5f995ef79e67dc4b083f8cfabe7a90798 -> trunk/fab85fc5f995ef79e67dc4b083f8cfabe7a90798 2025-07-17T08:34:52.3241604Z * [new tag] trunk/fac0cc16efd342637a3d1953caf2ecfa8a89947c -> trunk/fac0cc16efd342637a3d1953caf2ecfa8a89947c 2025-07-17T08:34:52.3243922Z * [new tag] trunk/fadc936fad0793e931ed2eb89577e1d10d212f71 -> trunk/fadc936fad0793e931ed2eb89577e1d10d212f71 2025-07-17T08:34:52.3246048Z * [new tag] trunk/fb45649df7267e97efffec8305cddcb23e97d53f -> trunk/fb45649df7267e97efffec8305cddcb23e97d53f 2025-07-17T08:34:52.3248380Z * [new tag] trunk/fb462cec8d8674ad547c55dbe90710bde1dc2019 -> trunk/fb462cec8d8674ad547c55dbe90710bde1dc2019 2025-07-17T08:34:52.3251718Z * [new tag] trunk/fb731fe371cb1b5bf95de84b19c213590526acb2 -> trunk/fb731fe371cb1b5bf95de84b19c213590526acb2 2025-07-17T08:34:52.3254091Z * [new tag] trunk/fb75dea2c1b93c78dccf08d5fd5e20b362ecd405 -> trunk/fb75dea2c1b93c78dccf08d5fd5e20b362ecd405 2025-07-17T08:34:52.3256598Z * [new tag] trunk/fb9a5d248f36ddce041025c8fc5be0d8bee454b0 -> trunk/fb9a5d248f36ddce041025c8fc5be0d8bee454b0 2025-07-17T08:34:52.3258882Z * [new tag] trunk/fbbab794ef6befbe7ffab3648e9b1f2042942fed -> trunk/fbbab794ef6befbe7ffab3648e9b1f2042942fed 2025-07-17T08:34:52.3261178Z * [new tag] trunk/fbd88ae2b5f444cb4b419c415258caeaef4b7325 -> trunk/fbd88ae2b5f444cb4b419c415258caeaef4b7325 2025-07-17T08:34:52.3263453Z * [new tag] trunk/fc0376e8b162d8fcf4375ae30566139eae7d48ed -> trunk/fc0376e8b162d8fcf4375ae30566139eae7d48ed 2025-07-17T08:34:52.3265791Z * [new tag] trunk/fc10d4b1d649f1460b587f3ab45ad4403d84518c -> trunk/fc10d4b1d649f1460b587f3ab45ad4403d84518c 2025-07-17T08:34:52.3268341Z * [new tag] trunk/fc177801afcbf12543b5e193e736b155f79d4ad3 -> trunk/fc177801afcbf12543b5e193e736b155f79d4ad3 2025-07-17T08:34:52.3270623Z * [new tag] trunk/fc5ae1229341ca2841a977197c644733c6a5c789 -> trunk/fc5ae1229341ca2841a977197c644733c6a5c789 2025-07-17T08:34:52.3272892Z * [new tag] trunk/fc772692622e89335dc28638c512bb33dbed1729 -> trunk/fc772692622e89335dc28638c512bb33dbed1729 2025-07-17T08:34:52.3275588Z * [new tag] trunk/fca7013f85d4e7ac07bdd461f490aa159feac6a0 -> trunk/fca7013f85d4e7ac07bdd461f490aa159feac6a0 2025-07-17T08:34:52.3277920Z * [new tag] trunk/fcbf7c749a839cc817927ceba8ea2887cc28dbf5 -> trunk/fcbf7c749a839cc817927ceba8ea2887cc28dbf5 2025-07-17T08:34:52.3280062Z * [new tag] trunk/fcc682be4bda58894a15fee1d9041c6043fea66f -> trunk/fcc682be4bda58894a15fee1d9041c6043fea66f 2025-07-17T08:34:52.3282520Z * [new tag] trunk/fd4bb29410c035b31ca55262c3012cadb1194aae -> trunk/fd4bb29410c035b31ca55262c3012cadb1194aae 2025-07-17T08:34:52.3284834Z * [new tag] trunk/fd4f704905f95b46c4c8fb4108461ff56ce750c5 -> trunk/fd4f704905f95b46c4c8fb4108461ff56ce750c5 2025-07-17T08:34:52.3287097Z * [new tag] trunk/fd8ea3c8a35e5b4aea3ebd712f2c7ee28c5a3655 -> trunk/fd8ea3c8a35e5b4aea3ebd712f2c7ee28c5a3655 2025-07-17T08:34:52.3289494Z * [new tag] trunk/fdc5b42a8fa2b15526c81f2c83f614e086056320 -> trunk/fdc5b42a8fa2b15526c81f2c83f614e086056320 2025-07-17T08:34:52.3291618Z * [new tag] trunk/fdf5d97fa8393f56aea2779877efd8a264ad5811 -> trunk/fdf5d97fa8393f56aea2779877efd8a264ad5811 2025-07-17T08:34:52.3294240Z * [new tag] trunk/fe1f1a38dff7c59ec0ec5b05fef058629845d061 -> trunk/fe1f1a38dff7c59ec0ec5b05fef058629845d061 2025-07-17T08:34:52.3296503Z * [new tag] trunk/fec571cfd458c4644a2f483dbf9f5480feca7939 -> trunk/fec571cfd458c4644a2f483dbf9f5480feca7939 2025-07-17T08:34:52.3298791Z * [new tag] trunk/fec8af8b98f5e17740ff947e9f8a1c447a497413 -> trunk/fec8af8b98f5e17740ff947e9f8a1c447a497413 2025-07-17T08:34:52.3300920Z * [new tag] trunk/fedbd1a48e1e474cf9da5637aae89b5bc4c20626 -> trunk/fedbd1a48e1e474cf9da5637aae89b5bc4c20626 2025-07-17T08:34:52.3303083Z * [new tag] trunk/fee2377f9ea62223f69ea9904c5e25ccb2af5961 -> trunk/fee2377f9ea62223f69ea9904c5e25ccb2af5961 2025-07-17T08:34:52.3305212Z * [new tag] trunk/feea575082439a0496dd404a4925b7d551039065 -> trunk/feea575082439a0496dd404a4925b7d551039065 2025-07-17T08:34:52.3307561Z * [new tag] trunk/ff611d971fe5362a71c15109cf020d30e6c4b2b9 -> trunk/ff611d971fe5362a71c15109cf020d30e6c4b2b9 2025-07-17T08:34:52.3309893Z * [new tag] trunk/ff7dd1776f9cb8448336338b19b9b53073f2fbda -> trunk/ff7dd1776f9cb8448336338b19b9b53073f2fbda 2025-07-17T08:34:52.3312298Z * [new tag] trunk/ff8b53c056e6556187690a37c944c92feb964d2d -> trunk/ff8b53c056e6556187690a37c944c92feb964d2d 2025-07-17T08:34:52.3314695Z * [new tag] trunk/ffac0de07e0173a073c92c157f43c515009c0de1 -> trunk/ffac0de07e0173a073c92c157f43c515009c0de1 2025-07-17T08:34:52.3317167Z * [new tag] trunk/ffaed8c569406839335bf46dafc4c3e8871e4b8a -> trunk/ffaed8c569406839335bf46dafc4c3e8871e4b8a 2025-07-17T08:34:52.3319501Z * [new tag] trunk/ffc6cbfaf78ca219092ce64dcf113377ae698300 -> trunk/ffc6cbfaf78ca219092ce64dcf113377ae698300 2025-07-17T08:34:52.3321821Z * [new tag] v0.1.1 -> v0.1.1 2025-07-17T08:34:52.3324175Z * [new tag] v0.1.10 -> v0.1.10 2025-07-17T08:34:52.3326451Z * [new tag] v0.1.11 -> v0.1.11 2025-07-17T08:34:52.3328803Z * [new tag] v0.1.12 -> v0.1.12 2025-07-17T08:34:52.3331043Z * [new tag] v0.1.2 -> v0.1.2 2025-07-17T08:34:52.3333300Z * [new tag] v0.1.3 -> v0.1.3 2025-07-17T08:34:52.3335669Z * [new tag] v0.1.4 -> v0.1.4 2025-07-17T08:34:52.3338004Z * [new tag] v0.1.5 -> v0.1.5 2025-07-17T08:34:52.3340343Z * [new tag] v0.1.6 -> v0.1.6 2025-07-17T08:34:52.3342713Z * [new tag] v0.1.7 -> v0.1.7 2025-07-17T08:34:52.3344945Z * [new tag] v0.1.8 -> v0.1.8 2025-07-17T08:34:52.3347255Z * [new tag] v0.1.9 -> v0.1.9 2025-07-17T08:34:52.3349608Z * [new tag] v0.2.0 -> v0.2.0 2025-07-17T08:34:52.3351968Z * [new tag] v0.3.0 -> v0.3.0 2025-07-17T08:34:52.3354448Z * [new tag] v0.3.1 -> v0.3.1 2025-07-17T08:34:52.3357017Z * [new tag] v0.4.0 -> v0.4.0 2025-07-17T08:34:52.3359326Z * [new tag] v0.4.1 -> v0.4.1 2025-07-17T08:34:52.3361615Z * [new tag] v1.0.0 -> v1.0.0 2025-07-17T08:34:52.3364061Z * [new tag] v1.0.0a0 -> v1.0.0a0 2025-07-17T08:34:52.3366393Z * [new tag] v1.0.1 -> v1.0.1 2025-07-17T08:34:52.3368937Z * [new tag] v1.0rc0 -> v1.0rc0 2025-07-17T08:34:52.3370889Z * [new tag] v1.0rc1 -> v1.0rc1 2025-07-17T08:34:52.3373243Z * [new tag] v1.1.0 -> v1.1.0 2025-07-17T08:34:52.3375551Z * [new tag] v1.1.0a0 -> v1.1.0a0 2025-07-17T08:34:52.3378087Z * [new tag] v1.10.0 -> v1.10.0 2025-07-17T08:34:52.3380572Z * [new tag] v1.10.0-rc1 -> v1.10.0-rc1 2025-07-17T08:34:52.3382947Z * [new tag] v1.10.0-rc2 -> v1.10.0-rc2 2025-07-17T08:34:52.3385093Z * [new tag] v1.10.0-rc3 -> v1.10.0-rc3 2025-07-17T08:34:52.3387536Z * [new tag] v1.10.1 -> v1.10.1 2025-07-17T08:34:52.3389696Z * [new tag] v1.10.1-rc1 -> v1.10.1-rc1 2025-07-17T08:34:52.3391809Z * [new tag] v1.10.2 -> v1.10.2 2025-07-17T08:34:52.3394025Z * [new tag] v1.10.2-rc1 -> v1.10.2-rc1 2025-07-17T08:34:52.3396569Z * [new tag] v1.11.0 -> v1.11.0 2025-07-17T08:34:52.3398915Z * [new tag] v1.11.0-rc1 -> v1.11.0-rc1 2025-07-17T08:34:52.3401360Z * [new tag] v1.11.0-rc2 -> v1.11.0-rc2 2025-07-17T08:34:52.3403761Z * [new tag] v1.11.0-rc3 -> v1.11.0-rc3 2025-07-17T08:34:52.3406179Z * [new tag] v1.11.0-rc4 -> v1.11.0-rc4 2025-07-17T08:34:52.3408606Z * [new tag] v1.11.0-rc5 -> v1.11.0-rc5 2025-07-17T08:34:52.3410733Z * [new tag] v1.11.0-rc6 -> v1.11.0-rc6 2025-07-17T08:34:52.3412932Z * [new tag] v1.11.0-rc7 -> v1.11.0-rc7 2025-07-17T08:34:52.3415392Z * [new tag] v1.12.0 -> v1.12.0 2025-07-17T08:34:52.3417730Z * [new tag] v1.12.0-rc1 -> v1.12.0-rc1 2025-07-17T08:34:52.3420143Z * [new tag] v1.12.0-rc2 -> v1.12.0-rc2 2025-07-17T08:34:52.3422562Z * [new tag] v1.12.0-rc3 -> v1.12.0-rc3 2025-07-17T08:34:52.3424993Z * [new tag] v1.12.0-rc4 -> v1.12.0-rc4 2025-07-17T08:34:52.3427270Z * [new tag] v1.12.0-rc5 -> v1.12.0-rc5 2025-07-17T08:34:52.3429705Z * [new tag] v1.12.0-rc6 -> v1.12.0-rc6 2025-07-17T08:34:52.3431967Z * [new tag] v1.12.0-rc7 -> v1.12.0-rc7 2025-07-17T08:34:52.3434158Z * [new tag] v1.12.0-rc8 -> v1.12.0-rc8 2025-07-17T08:34:52.3436841Z * [new tag] v1.12.1 -> v1.12.1 2025-07-17T08:34:52.3439324Z * [new tag] v1.12.1-rc1 -> v1.12.1-rc1 2025-07-17T08:34:52.3441693Z * [new tag] v1.12.1-rc2 -> v1.12.1-rc2 2025-07-17T08:34:52.3444103Z * [new tag] v1.12.1-rc3 -> v1.12.1-rc3 2025-07-17T08:34:52.3446500Z * [new tag] v1.12.1-rc4 -> v1.12.1-rc4 2025-07-17T08:34:52.3448639Z * [new tag] v1.12.1-rc5 -> v1.12.1-rc5 2025-07-17T08:34:52.3451070Z * [new tag] v1.13.0 -> v1.13.0 2025-07-17T08:34:52.3453376Z * [new tag] v1.13.0-rc1 -> v1.13.0-rc1 2025-07-17T08:34:52.3455707Z * [new tag] v1.13.0-rc2 -> v1.13.0-rc2 2025-07-17T08:34:52.3458053Z * [new tag] v1.13.0-rc3 -> v1.13.0-rc3 2025-07-17T08:34:52.3460430Z * [new tag] v1.13.0-rc4 -> v1.13.0-rc4 2025-07-17T08:34:52.3462550Z * [new tag] v1.13.0-rc5 -> v1.13.0-rc5 2025-07-17T08:34:52.3464860Z * [new tag] v1.13.0-rc6 -> v1.13.0-rc6 2025-07-17T08:34:52.3467111Z * [new tag] v1.13.1 -> v1.13.1 2025-07-17T08:34:52.3469269Z * [new tag] v1.13.1-rc1 -> v1.13.1-rc1 2025-07-17T08:34:52.3471728Z * [new tag] v1.2.0 -> v1.2.0 2025-07-17T08:34:52.3474161Z * [new tag] v1.2.0a0 -> v1.2.0a0 2025-07-17T08:34:52.3476579Z * [new tag] v1.3.0 -> v1.3.0 2025-07-17T08:34:52.3479017Z * [new tag] v1.3.0a0 -> v1.3.0a0 2025-07-17T08:34:52.3481210Z * [new tag] v1.3.1 -> v1.3.1 2025-07-17T08:34:52.3483600Z * [new tag] v1.4.0 -> v1.4.0 2025-07-17T08:34:52.3485951Z * [new tag] v1.4.0a0 -> v1.4.0a0 2025-07-17T08:34:52.3488105Z * [new tag] v1.4.1 -> v1.4.1 2025-07-17T08:34:52.3490510Z * [new tag] v1.5.0 -> v1.5.0 2025-07-17T08:34:52.3492916Z * [new tag] v1.5.0-rc1 -> v1.5.0-rc1 2025-07-17T08:34:52.3495279Z * [new tag] v1.5.0-rc2 -> v1.5.0-rc2 2025-07-17T08:34:52.3497731Z * [new tag] v1.5.0-rc3 -> v1.5.0-rc3 2025-07-17T08:34:52.3499985Z * [new tag] v1.5.0-rc4 -> v1.5.0-rc4 2025-07-17T08:34:52.3502101Z * [new tag] v1.5.0-rc5 -> v1.5.0-rc5 2025-07-17T08:34:52.3504552Z * [new tag] v1.5.1 -> v1.5.1 2025-07-17T08:34:52.3506794Z * [new tag] v1.5.1-rc1 -> v1.5.1-rc1 2025-07-17T08:34:52.3508967Z * [new tag] v1.6.0 -> v1.6.0 2025-07-17T08:34:52.3511293Z * [new tag] v1.6.0-rc1 -> v1.6.0-rc1 2025-07-17T08:34:52.3513723Z * [new tag] v1.6.0-rc2 -> v1.6.0-rc2 2025-07-17T08:34:52.3516383Z * [new tag] v1.6.0-rc3 -> v1.6.0-rc3 2025-07-17T08:34:52.3518811Z * [new tag] v1.6.0-rc4 -> v1.6.0-rc4 2025-07-17T08:34:52.3521152Z * [new tag] v1.6.0-rc5 -> v1.6.0-rc5 2025-07-17T08:34:52.3523458Z * [new tag] v1.6.0-rc6 -> v1.6.0-rc6 2025-07-17T08:34:52.3525609Z * [new tag] v1.6.0-rc7 -> v1.6.0-rc7 2025-07-17T08:34:52.3528074Z * [new tag] v1.7.0 -> v1.7.0 2025-07-17T08:34:52.3530411Z * [new tag] v1.7.0-rc1 -> v1.7.0-rc1 2025-07-17T08:34:52.3532837Z * [new tag] v1.7.0-rc2 -> v1.7.0-rc2 2025-07-17T08:34:52.3535195Z * [new tag] v1.7.0-rc3 -> v1.7.0-rc3 2025-07-17T08:34:52.3537343Z * [new tag] v1.7.0-rc4 -> v1.7.0-rc4 2025-07-17T08:34:52.3539728Z * [new tag] v1.7.1 -> v1.7.1 2025-07-17T08:34:52.3542191Z * [new tag] v1.7.1-rc1 -> v1.7.1-rc1 2025-07-17T08:34:52.3544584Z * [new tag] v1.7.1-rc2 -> v1.7.1-rc2 2025-07-17T08:34:52.3546724Z * [new tag] v1.7.1-rc3 -> v1.7.1-rc3 2025-07-17T08:34:52.3549121Z * [new tag] v1.8.0 -> v1.8.0 2025-07-17T08:34:52.3551259Z * [new tag] v1.8.0-rc1 -> v1.8.0-rc1 2025-07-17T08:34:52.3553650Z * [new tag] v1.8.0-rc2 -> v1.8.0-rc2 2025-07-17T08:34:52.3556118Z * [new tag] v1.8.0-rc3 -> v1.8.0-rc3 2025-07-17T08:34:52.3558429Z * [new tag] v1.8.0-rc4 -> v1.8.0-rc4 2025-07-17T08:34:52.3560732Z * [new tag] v1.8.0-rc5 -> v1.8.0-rc5 2025-07-17T08:34:52.3562740Z * [new tag] v1.8.1 -> v1.8.1 2025-07-17T08:34:52.3565104Z * [new tag] v1.8.1-rc1 -> v1.8.1-rc1 2025-07-17T08:34:52.3567236Z * [new tag] v1.8.1-rc2 -> v1.8.1-rc2 2025-07-17T08:34:52.3569390Z * [new tag] v1.8.1-rc3 -> v1.8.1-rc3 2025-07-17T08:34:52.3572145Z * [new tag] v1.8.2 -> v1.8.2 2025-07-17T08:34:52.3574280Z * [new tag] v1.8.2-rc1 -> v1.8.2-rc1 2025-07-17T08:34:52.3576649Z * [new tag] v1.9.0 -> v1.9.0 2025-07-17T08:34:52.3578991Z * [new tag] v1.9.0-rc1 -> v1.9.0-rc1 2025-07-17T08:34:52.3581538Z * [new tag] v1.9.0-rc2 -> v1.9.0-rc2 2025-07-17T08:34:52.3583934Z * [new tag] v1.9.0-rc3 -> v1.9.0-rc3 2025-07-17T08:34:52.3586046Z * [new tag] v1.9.0-rc4 -> v1.9.0-rc4 2025-07-17T08:34:52.3588425Z * [new tag] v1.9.1 -> v1.9.1 2025-07-17T08:34:52.3590927Z * [new tag] v1.9.1-rc1 -> v1.9.1-rc1 2025-07-17T08:34:52.3593077Z * [new tag] v1.9.1-rc2 -> v1.9.1-rc2 2025-07-17T08:34:52.3595618Z * [new tag] v2.0.0 -> v2.0.0 2025-07-17T08:34:52.3597914Z * [new tag] v2.0.0-rc1 -> v2.0.0-rc1 2025-07-17T08:34:52.3600307Z * [new tag] v2.0.0-rc2 -> v2.0.0-rc2 2025-07-17T08:34:52.3602793Z * [new tag] v2.0.0-rc3 -> v2.0.0-rc3 2025-07-17T08:34:52.3605177Z * [new tag] v2.0.0-rc4 -> v2.0.0-rc4 2025-07-17T08:34:52.3607501Z * [new tag] v2.0.0-rc5 -> v2.0.0-rc5 2025-07-17T08:34:52.3609677Z * [new tag] v2.0.0-rc6 -> v2.0.0-rc6 2025-07-17T08:34:52.3612088Z * [new tag] v2.0.1 -> v2.0.1 2025-07-17T08:34:52.3614563Z * [new tag] v2.0.1-rc1 -> v2.0.1-rc1 2025-07-17T08:34:52.3616715Z * [new tag] v2.0.1-rc2 -> v2.0.1-rc2 2025-07-17T08:34:52.3618975Z * [new tag] v2.0.1-rc3 -> v2.0.1-rc3 2025-07-17T08:34:52.3621159Z * [new tag] v2.0.1-rc4 -> v2.0.1-rc4 2025-07-17T08:34:52.3623934Z * [new tag] v2.1.0 -> v2.1.0 2025-07-17T08:34:52.3626253Z * [new tag] v2.1.0-rc1 -> v2.1.0-rc1 2025-07-17T08:34:52.3628637Z * [new tag] v2.1.0-rc2 -> v2.1.0-rc2 2025-07-17T08:34:52.3631048Z * [new tag] v2.1.0-rc3 -> v2.1.0-rc3 2025-07-17T08:34:52.3633461Z * [new tag] v2.1.0-rc4 -> v2.1.0-rc4 2025-07-17T08:34:52.3635984Z * [new tag] v2.1.0-rc5 -> v2.1.0-rc5 2025-07-17T08:34:52.3638114Z * [new tag] v2.1.0-rc6 -> v2.1.0-rc6 2025-07-17T08:34:52.3640453Z * [new tag] v2.1.1 -> v2.1.1 2025-07-17T08:34:52.3642776Z * [new tag] v2.1.1-rc1 -> v2.1.1-rc1 2025-07-17T08:34:52.3645138Z * [new tag] v2.1.1-rc2 -> v2.1.1-rc2 2025-07-17T08:34:52.3647602Z * [new tag] v2.1.1-rc3 -> v2.1.1-rc3 2025-07-17T08:34:52.3650008Z * [new tag] v2.1.1-rc4 -> v2.1.1-rc4 2025-07-17T08:34:52.3652339Z * [new tag] v2.1.1-rc5 -> v2.1.1-rc5 2025-07-17T08:34:52.3654491Z * [new tag] v2.1.1-rc6 -> v2.1.1-rc6 2025-07-17T08:34:52.3656950Z * [new tag] v2.1.2 -> v2.1.2 2025-07-17T08:34:52.3659546Z * [new tag] v2.1.2-rc1 -> v2.1.2-rc1 2025-07-17T08:34:52.3661742Z * [new tag] v2.1.2-rc2 -> v2.1.2-rc2 2025-07-17T08:34:52.3663876Z * [new tag] v2.1.2-rc3 -> v2.1.2-rc3 2025-07-17T08:34:52.3666226Z * [new tag] v2.2.0 -> v2.2.0 2025-07-17T08:34:52.3668591Z * [new tag] v2.2.0-rc1 -> v2.2.0-rc1 2025-07-17T08:34:52.3670874Z * [new tag] v2.2.0-rc2 -> v2.2.0-rc2 2025-07-17T08:34:52.3673162Z * [new tag] v2.2.0-rc3 -> v2.2.0-rc3 2025-07-17T08:34:52.3676955Z * [new tag] v2.2.0-rc4 -> v2.2.0-rc4 2025-07-17T08:34:52.3679337Z * [new tag] v2.2.0-rc5 -> v2.2.0-rc5 2025-07-17T08:34:52.3681740Z * [new tag] v2.2.0-rc6 -> v2.2.0-rc6 2025-07-17T08:34:52.3683864Z * [new tag] v2.2.0-rc7 -> v2.2.0-rc7 2025-07-17T08:34:52.3686009Z * [new tag] v2.2.0-rc8 -> v2.2.0-rc8 2025-07-17T08:34:52.3688322Z * [new tag] v2.2.1 -> v2.2.1 2025-07-17T08:34:52.3690688Z * [new tag] v2.2.1-rc1 -> v2.2.1-rc1 2025-07-17T08:34:52.3692830Z * [new tag] v2.2.1-rc2 -> v2.2.1-rc2 2025-07-17T08:34:52.3694968Z * [new tag] v2.2.1-rc3 -> v2.2.1-rc3 2025-07-17T08:34:52.3697119Z * [new tag] v2.2.2 -> v2.2.2 2025-07-17T08:34:52.3699525Z * [new tag] v2.2.2-rc1 -> v2.2.2-rc1 2025-07-17T08:34:52.3701631Z * [new tag] v2.2.2-rc2 -> v2.2.2-rc2 2025-07-17T08:34:52.3703759Z * [new tag] v2.2.2-rc3 -> v2.2.2-rc3 2025-07-17T08:34:52.3706160Z * [new tag] v2.3.0 -> v2.3.0 2025-07-17T08:34:52.3708474Z * [new tag] v2.3.0-rc1 -> v2.3.0-rc1 2025-07-17T08:34:52.3710903Z * [new tag] v2.3.0-rc10 -> v2.3.0-rc10 2025-07-17T08:34:52.3713310Z * [new tag] v2.3.0-rc11 -> v2.3.0-rc11 2025-07-17T08:34:52.3715707Z * [new tag] v2.3.0-rc12 -> v2.3.0-rc12 2025-07-17T08:34:52.3718041Z * [new tag] v2.3.0-rc2 -> v2.3.0-rc2 2025-07-17T08:34:52.3720394Z * [new tag] v2.3.0-rc3 -> v2.3.0-rc3 2025-07-17T08:34:52.3722751Z * [new tag] v2.3.0-rc4 -> v2.3.0-rc4 2025-07-17T08:34:52.3725034Z * [new tag] v2.3.0-rc5 -> v2.3.0-rc5 2025-07-17T08:34:52.3727148Z * [new tag] v2.3.0-rc6 -> v2.3.0-rc6 2025-07-17T08:34:52.3729527Z * [new tag] v2.3.0-rc7 -> v2.3.0-rc7 2025-07-17T08:34:52.3731971Z * [new tag] v2.3.0-rc8 -> v2.3.0-rc8 2025-07-17T08:34:52.3734086Z * [new tag] v2.3.0-rc9 -> v2.3.0-rc9 2025-07-17T08:34:52.3736207Z * [new tag] v2.3.1 -> v2.3.1 2025-07-17T08:34:52.3738563Z * [new tag] v2.3.1-rc1 -> v2.3.1-rc1 2025-07-17T08:34:52.3740988Z * [new tag] v2.3.1-rc2 -> v2.3.1-rc2 2025-07-17T08:34:52.3744034Z * [new tag] v2.3.1-rc3 -> v2.3.1-rc3 2025-07-17T08:34:52.3746417Z * [new tag] v2.4.0 -> v2.4.0 2025-07-17T08:34:52.3748798Z * [new tag] v2.4.0-rc1 -> v2.4.0-rc1 2025-07-17T08:34:52.3751120Z * [new tag] v2.4.0-rc2 -> v2.4.0-rc2 2025-07-17T08:34:52.3753400Z * [new tag] v2.4.0-rc3 -> v2.4.0-rc3 2025-07-17T08:34:52.3756183Z * [new tag] v2.4.0-rc4 -> v2.4.0-rc4 2025-07-17T08:34:52.3758341Z * [new tag] v2.4.0-rc5 -> v2.4.0-rc5 2025-07-17T08:34:52.3760664Z * [new tag] v2.4.0-rc6 -> v2.4.0-rc6 2025-07-17T08:34:52.3763036Z * [new tag] v2.4.0-rc7 -> v2.4.0-rc7 2025-07-17T08:34:52.3765368Z * [new tag] v2.4.0-rc8 -> v2.4.0-rc8 2025-07-17T08:34:52.3767835Z * [new tag] v2.4.0-rc9 -> v2.4.0-rc9 2025-07-17T08:34:52.3769988Z * [new tag] v2.4.1 -> v2.4.1 2025-07-17T08:34:52.3772443Z * [new tag] v2.4.1-rc1 -> v2.4.1-rc1 2025-07-17T08:34:52.3774776Z * [new tag] v2.4.1-rc2 -> v2.4.1-rc2 2025-07-17T08:34:52.3777118Z * [new tag] v2.4.1-rc3 -> v2.4.1-rc3 2025-07-17T08:34:52.3779515Z * [new tag] v2.5.0 -> v2.5.0 2025-07-17T08:34:52.3781852Z * [new tag] v2.5.0-rc1 -> v2.5.0-rc1 2025-07-17T08:34:52.3784007Z * [new tag] v2.5.0-rc10 -> v2.5.0-rc10 2025-07-17T08:34:52.3786349Z * [new tag] v2.5.0-rc2 -> v2.5.0-rc2 2025-07-17T08:34:52.3788737Z * [new tag] v2.5.0-rc3 -> v2.5.0-rc3 2025-07-17T08:34:52.3791066Z * [new tag] v2.5.0-rc4 -> v2.5.0-rc4 2025-07-17T08:34:52.3793435Z * [new tag] v2.5.0-rc5 -> v2.5.0-rc5 2025-07-17T08:34:52.3795960Z * [new tag] v2.5.0-rc6 -> v2.5.0-rc6 2025-07-17T08:34:52.3798272Z * [new tag] v2.5.0-rc7 -> v2.5.0-rc7 2025-07-17T08:34:52.3800727Z * [new tag] v2.5.0-rc8 -> v2.5.0-rc8 2025-07-17T08:34:52.3803121Z * [new tag] v2.5.0-rc9 -> v2.5.0-rc9 2025-07-17T08:34:52.3805255Z * [new tag] v2.5.1 -> v2.5.1 2025-07-17T08:34:52.3807537Z * [new tag] v2.5.1-rc1 -> v2.5.1-rc1 2025-07-17T08:34:52.3809667Z * [new tag] v2.6.0 -> v2.6.0 2025-07-17T08:34:52.3812108Z * [new tag] v2.6.0-rc1 -> v2.6.0-rc1 2025-07-17T08:34:52.3814482Z * [new tag] v2.6.0-rc2 -> v2.6.0-rc2 2025-07-17T08:34:52.3816846Z * [new tag] v2.6.0-rc3 -> v2.6.0-rc3 2025-07-17T08:34:52.3819129Z * [new tag] v2.6.0-rc4 -> v2.6.0-rc4 2025-07-17T08:34:52.3821726Z * [new tag] v2.6.0-rc5 -> v2.6.0-rc5 2025-07-17T08:34:52.3824149Z * [new tag] v2.6.0-rc6 -> v2.6.0-rc6 2025-07-17T08:34:52.3826573Z * [new tag] v2.6.0-rc7 -> v2.6.0-rc7 2025-07-17T08:34:52.3828971Z * [new tag] v2.6.0-rc8 -> v2.6.0-rc8 2025-07-17T08:34:52.3831338Z * [new tag] v2.6.0-rc9 -> v2.6.0-rc9 2025-07-17T08:34:52.3833833Z * [new tag] v2.7.0 -> v2.7.0 2025-07-17T08:34:52.3836276Z * [new tag] v2.7.0-rc1 -> v2.7.0-rc1 2025-07-17T08:34:52.3838420Z * [new tag] v2.7.0-rc10 -> v2.7.0-rc10 2025-07-17T08:34:52.3840871Z * [new tag] v2.7.0-rc2 -> v2.7.0-rc2 2025-07-17T08:34:52.3843180Z * [new tag] v2.7.0-rc3 -> v2.7.0-rc3 2025-07-17T08:34:52.3845607Z * [new tag] v2.7.0-rc4 -> v2.7.0-rc4 2025-07-17T08:34:52.3847939Z * [new tag] v2.7.0-rc5 -> v2.7.0-rc5 2025-07-17T08:34:52.3850253Z * [new tag] v2.7.0-rc6 -> v2.7.0-rc6 2025-07-17T08:34:52.3852848Z * [new tag] v2.7.0-rc7 -> v2.7.0-rc7 2025-07-17T08:34:52.3854969Z * [new tag] v2.7.0-rc8 -> v2.7.0-rc8 2025-07-17T08:34:52.3857378Z * [new tag] v2.7.0-rc9 -> v2.7.0-rc9 2025-07-17T08:34:52.3859497Z * [new tag] v2.7.1 -> v2.7.1 2025-07-17T08:34:52.3861833Z * [new tag] v2.7.1-rc1 -> v2.7.1-rc1 2025-07-17T08:34:52.3864204Z * [new tag] v2.7.1-rc2 -> v2.7.1-rc2 2025-07-17T08:34:52.3866558Z * [new tag] v2.7.1-rc3 -> v2.7.1-rc3 2025-07-17T08:34:52.3868874Z * [new tag] v2.7.1-rc4 -> v2.7.1-rc4 2025-07-17T08:34:52.3871048Z * [new tag] v2.7.1-rc5 -> v2.7.1-rc5 2025-07-17T08:34:52.3873494Z * [new tag] v2.8.0-rc1 -> v2.8.0-rc1 2025-07-17T08:34:52.3875877Z * [new tag] v2.8.0-rc2 -> v2.8.0-rc2 2025-07-17T08:34:52.3878284Z * [new tag] v2.8.0-rc3 -> v2.8.0-rc3 2025-07-17T08:34:52.3880645Z * [new tag] v2.8.0-rc4 -> v2.8.0-rc4 2025-07-17T08:34:52.3883106Z * [new tag] v2.8.0-rc5 -> v2.8.0-rc5 2025-07-17T08:34:52.3885405Z * [new tag] whc_flight_1 -> whc_flight_1 2025-07-17T08:34:52.3887567Z * [new tag] whc_flight_2 -> whc_flight_2 2025-07-17T08:34:52.3889790Z * [new tag] whc_flight_4 -> whc_flight_4 2025-07-17T08:34:52.6096147Z [command]/usr/bin/git rev-parse --verify --quiet a38f433be2e94a64b095a44ba39879d02d0c2316^{object} 2025-07-17T08:34:52.6128862Z a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:34:52.6133686Z ##[endgroup] 2025-07-17T08:34:52.6134066Z ##[group]Determining the checkout info 2025-07-17T08:34:52.6134419Z ##[endgroup] 2025-07-17T08:34:52.6138052Z [command]/usr/bin/git sparse-checkout disable 2025-07-17T08:34:52.6189136Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-07-17T08:34:52.6224699Z ##[group]Checking out the ref 2025-07-17T08:34:52.6227547Z [command]/usr/bin/git checkout --progress --force a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:34:53.6431851Z Updating files: 42% (8409/19589) 2025-07-17T08:34:53.6535427Z Updating files: 43% (8424/19589) 2025-07-17T08:34:53.6642310Z Updating files: 44% (8620/19589) 2025-07-17T08:34:53.6748747Z Updating files: 45% (8816/19589) 2025-07-17T08:34:53.6851962Z Updating files: 46% (9011/19589) 2025-07-17T08:34:53.6954613Z Updating files: 47% (9207/19589) 2025-07-17T08:34:53.7061175Z Updating files: 48% (9403/19589) 2025-07-17T08:34:53.7166094Z Updating files: 49% (9599/19589) 2025-07-17T08:34:53.7271415Z Updating files: 50% (9795/19589) 2025-07-17T08:34:53.7376034Z Updating files: 51% (9991/19589) 2025-07-17T08:34:53.7482202Z Updating files: 52% (10187/19589) 2025-07-17T08:34:53.7588899Z Updating files: 53% (10383/19589) 2025-07-17T08:34:53.7692377Z Updating files: 54% (10579/19589) 2025-07-17T08:34:53.7798189Z Updating files: 55% (10774/19589) 2025-07-17T08:34:53.7904910Z Updating files: 56% (10970/19589) 2025-07-17T08:34:53.8014806Z Updating files: 57% (11166/19589) 2025-07-17T08:34:53.8119582Z Updating files: 58% (11362/19589) 2025-07-17T08:34:53.8223966Z Updating files: 59% (11558/19589) 2025-07-17T08:34:53.8330403Z Updating files: 60% (11754/19589) 2025-07-17T08:34:53.8436732Z Updating files: 61% (11950/19589) 2025-07-17T08:34:53.8542592Z Updating files: 62% (12146/19589) 2025-07-17T08:34:53.8650230Z Updating files: 63% (12342/19589) 2025-07-17T08:34:53.8758710Z Updating files: 64% (12537/19589) 2025-07-17T08:34:53.8868901Z Updating files: 65% (12733/19589) 2025-07-17T08:34:53.8976916Z Updating files: 66% (12929/19589) 2025-07-17T08:34:53.9082816Z Updating files: 67% (13125/19589) 2025-07-17T08:34:53.9185877Z Updating files: 68% (13321/19589) 2025-07-17T08:34:53.9290353Z Updating files: 69% (13517/19589) 2025-07-17T08:34:53.9562813Z Updating files: 70% (13713/19589) 2025-07-17T08:34:53.9865521Z Updating files: 71% (13909/19589) 2025-07-17T08:34:53.9968794Z Updating files: 72% (14105/19589) 2025-07-17T08:34:54.0126264Z Updating files: 73% (14300/19589) 2025-07-17T08:34:54.0459976Z Updating files: 74% (14496/19589) 2025-07-17T08:34:54.0774108Z Updating files: 75% (14692/19589) 2025-07-17T08:34:54.1347788Z Updating files: 76% (14888/19589) 2025-07-17T08:34:54.1569792Z Updating files: 77% (15084/19589) 2025-07-17T08:34:54.1764105Z Updating files: 78% (15280/19589) 2025-07-17T08:34:54.2100468Z Updating files: 79% (15476/19589) 2025-07-17T08:34:54.2427032Z Updating files: 80% (15672/19589) 2025-07-17T08:34:54.2809145Z Updating files: 81% (15868/19589) 2025-07-17T08:34:54.3181292Z Updating files: 82% (16063/19589) 2025-07-17T08:34:54.3442381Z Updating files: 83% (16259/19589) 2025-07-17T08:34:54.3631311Z Updating files: 84% (16455/19589) 2025-07-17T08:34:54.3864341Z Updating files: 85% (16651/19589) 2025-07-17T08:34:54.4109967Z Updating files: 86% (16847/19589) 2025-07-17T08:34:54.4330276Z Updating files: 87% (17043/19589) 2025-07-17T08:34:54.4547066Z Updating files: 88% (17239/19589) 2025-07-17T08:34:54.4743102Z Updating files: 89% (17435/19589) 2025-07-17T08:34:54.4990100Z Updating files: 90% (17631/19589) 2025-07-17T08:34:54.5210480Z Updating files: 91% (17826/19589) 2025-07-17T08:34:54.5419910Z Updating files: 92% (18022/19589) 2025-07-17T08:34:54.5670832Z Updating files: 93% (18218/19589) 2025-07-17T08:34:54.5941625Z Updating files: 94% (18414/19589) 2025-07-17T08:34:54.6236669Z Updating files: 95% (18610/19589) 2025-07-17T08:34:54.6423093Z Updating files: 96% (18806/19589) 2025-07-17T08:34:54.6491788Z Updating files: 96% (18950/19589) 2025-07-17T08:34:54.6748145Z Updating files: 97% (19002/19589) 2025-07-17T08:34:54.7121517Z Updating files: 98% (19198/19589) 2025-07-17T08:34:54.7368700Z Updating files: 99% (19394/19589) 2025-07-17T08:34:54.7370094Z Updating files: 100% (19589/19589) 2025-07-17T08:34:54.7370409Z Updating files: 100% (19589/19589), done. 2025-07-17T08:34:54.7568725Z Note: switching to 'a38f433be2e94a64b095a44ba39879d02d0c2316'. 2025-07-17T08:34:54.7568990Z 2025-07-17T08:34:54.7569158Z You are in 'detached HEAD' state. You can look around, make experimental 2025-07-17T08:34:54.7569493Z changes and commit them, and you can discard any commits you make in this 2025-07-17T08:34:54.7569824Z state without impacting any branches by switching back to a branch. 2025-07-17T08:34:54.7570045Z 2025-07-17T08:34:54.7570172Z If you want to create a new branch to retain commits you create, you may 2025-07-17T08:34:54.7570506Z do so (now or later) by using -c with the switch command. Example: 2025-07-17T08:34:54.7570677Z 2025-07-17T08:34:54.7570769Z git switch -c 2025-07-17T08:34:54.7570898Z 2025-07-17T08:34:54.7570975Z Or undo this operation with: 2025-07-17T08:34:54.7571097Z 2025-07-17T08:34:54.7571182Z git switch - 2025-07-17T08:34:54.7571290Z 2025-07-17T08:34:54.7571443Z Turn off this advice by setting config variable advice.detachedHead to false 2025-07-17T08:34:54.7571647Z 2025-07-17T08:34:54.7571816Z HEAD is now at a38f433be2e [Docker builds] Move from Miniconda to Miniforge (#158370) 2025-07-17T08:34:54.7724158Z ##[endgroup] 2025-07-17T08:34:54.7724492Z ##[group]Setting up auth for fetching submodules 2025-07-17T08:34:54.7729534Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-07-17T08:34:54.7778604Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-07-17T08:34:54.7814985Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-07-17T08:34:54.7845070Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-07-17T08:34:54.7872891Z ##[endgroup] 2025-07-17T08:34:54.7873206Z ##[group]Fetching submodules 2025-07-17T08:34:54.7876216Z [command]/usr/bin/git submodule sync --recursive 2025-07-17T08:34:54.8250793Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-07-17T08:34:54.8555547Z Submodule 'android/libs/fbjni' (https://github.com/facebookincubator/fbjni.git) registered for path 'android/libs/fbjni' 2025-07-17T08:34:54.8560904Z Submodule 'third_party/NNPACK_deps/FP16' (https://github.com/Maratyszcza/FP16.git) registered for path 'third_party/FP16' 2025-07-17T08:34:54.8566926Z Submodule 'third_party/NNPACK_deps/FXdiv' (https://github.com/Maratyszcza/FXdiv.git) registered for path 'third_party/FXdiv' 2025-07-17T08:34:54.8573435Z Submodule 'third_party/NNPACK' (https://github.com/Maratyszcza/NNPACK.git) registered for path 'third_party/NNPACK' 2025-07-17T08:34:54.8579985Z Submodule 'third_party/NVTX' (https://github.com/NVIDIA/NVTX.git) registered for path 'third_party/NVTX' 2025-07-17T08:34:54.8586984Z Submodule 'third_party/VulkanMemoryAllocator' (https://github.com/GPUOpen-LibrariesAndSDKs/VulkanMemoryAllocator.git) registered for path 'third_party/VulkanMemoryAllocator' 2025-07-17T08:34:54.8593822Z Submodule 'third_party/XNNPACK' (https://github.com/google/XNNPACK.git) registered for path 'third_party/XNNPACK' 2025-07-17T08:34:54.8600776Z Submodule 'third_party/aiter' (https://github.com/ROCm/aiter.git) registered for path 'third_party/aiter' 2025-07-17T08:34:54.8607663Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/benchmark' 2025-07-17T08:34:54.8614762Z Submodule 'third_party/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/composable_kernel' 2025-07-17T08:34:54.8621882Z Submodule 'third_party/cpp-httplib' (https://github.com/yhirose/cpp-httplib.git) registered for path 'third_party/cpp-httplib' 2025-07-17T08:34:54.8629000Z Submodule 'third_party/cpuinfo' (https://github.com/pytorch/cpuinfo.git) registered for path 'third_party/cpuinfo' 2025-07-17T08:34:54.8636505Z Submodule 'third_party/cudnn_frontend' (https://github.com/NVIDIA/cudnn-frontend.git) registered for path 'third_party/cudnn_frontend' 2025-07-17T08:34:54.8643687Z Submodule 'third_party/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/cutlass' 2025-07-17T08:34:54.8650919Z Submodule 'third_party/fbgemm' (https://github.com/pytorch/fbgemm) registered for path 'third_party/fbgemm' 2025-07-17T08:34:54.8658335Z Submodule 'third_party/flash-attention' (https://github.com/Dao-AILab/flash-attention.git) registered for path 'third_party/flash-attention' 2025-07-17T08:34:54.8665729Z Submodule 'third_party/flatbuffers' (https://github.com/google/flatbuffers.git) registered for path 'third_party/flatbuffers' 2025-07-17T08:34:54.8673424Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/fmt' 2025-07-17T08:34:54.8681506Z Submodule 'third_party/gemmlowp/gemmlowp' (https://github.com/google/gemmlowp.git) registered for path 'third_party/gemmlowp/gemmlowp' 2025-07-17T08:34:54.8689098Z Submodule 'third_party/gloo' (https://github.com/pytorch/gloo) registered for path 'third_party/gloo' 2025-07-17T08:34:54.8696858Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/googletest' 2025-07-17T08:34:54.8704555Z Submodule 'third_party/ideep' (https://github.com/intel/ideep) registered for path 'third_party/ideep' 2025-07-17T08:34:54.8712621Z Submodule 'third_party/ittapi' (https://github.com/intel/ittapi.git) registered for path 'third_party/ittapi' 2025-07-17T08:34:54.8720653Z Submodule 'third_party/kineto' (https://github.com/pytorch/kineto) registered for path 'third_party/kineto' 2025-07-17T08:34:54.8728858Z Submodule 'third_party/kleidiai' (https://github.com/ARM-software/kleidiai.git) registered for path 'third_party/kleidiai' 2025-07-17T08:34:54.8736883Z Submodule 'third_party/mimalloc' (https://github.com/microsoft/mimalloc.git) registered for path 'third_party/mimalloc' 2025-07-17T08:34:54.8745330Z Submodule 'third_party/nlohmann' (https://github.com/nlohmann/json.git) registered for path 'third_party/nlohmann' 2025-07-17T08:34:54.8753212Z Submodule 'third_party/onnx' (https://github.com/onnx/onnx.git) registered for path 'third_party/onnx' 2025-07-17T08:34:54.8761835Z Submodule 'third_party/opentelemetry-cpp' (https://github.com/open-telemetry/opentelemetry-cpp.git) registered for path 'third_party/opentelemetry-cpp' 2025-07-17T08:34:54.8770069Z Submodule 'third_party/pocketfft' (https://github.com/mreineck/pocketfft) registered for path 'third_party/pocketfft' 2025-07-17T08:34:54.8778502Z Submodule 'third_party/protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'third_party/protobuf' 2025-07-17T08:34:54.8786957Z Submodule 'third_party/NNPACK_deps/psimd' (https://github.com/Maratyszcza/psimd.git) registered for path 'third_party/psimd' 2025-07-17T08:34:54.8795839Z Submodule 'third_party/NNPACK_deps/pthreadpool' (https://github.com/Maratyszcza/pthreadpool.git) registered for path 'third_party/pthreadpool' 2025-07-17T08:34:54.8804459Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/pybind11' 2025-07-17T08:34:54.8813219Z Submodule 'third_party/python-peachpy' (https://github.com/malfet/PeachPy.git) registered for path 'third_party/python-peachpy' 2025-07-17T08:34:54.8821958Z Submodule 'third_party/sleef' (https://github.com/shibatch/sleef) registered for path 'third_party/sleef' 2025-07-17T08:34:54.8830950Z Submodule 'third_party/tensorpipe' (https://github.com/pytorch/tensorpipe.git) registered for path 'third_party/tensorpipe' 2025-07-17T08:34:54.8874657Z Cloning into '/home/runner/_work/pytorch/pytorch/android/libs/fbjni'... 2025-07-17T08:34:59.3640259Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/pocketfft'... 2025-07-17T08:34:59.3640685Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/psimd'... 2025-07-17T08:34:59.3641028Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/pthreadpool'... 2025-07-17T08:34:59.3641411Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/python-peachpy'... 2025-07-17T08:34:59.3641753Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/ideep'... 2025-07-17T08:34:59.3642082Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/gemmlowp/gemmlowp'... 2025-07-17T08:34:59.3642415Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/gloo'... 2025-07-17T08:34:59.3642727Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/benchmark'... 2025-07-17T08:34:59.3643065Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/tensorpipe'... 2025-07-17T08:34:59.3643392Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/ittapi'... 2025-07-17T08:34:59.3643703Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kleidiai'... 2025-07-17T08:34:59.3644038Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/cpp-httplib'... 2025-07-17T08:34:59.3644389Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/flash-attention'... 2025-07-17T08:34:59.3644716Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/cpuinfo'... 2025-07-17T08:34:59.3645045Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/googletest'... 2025-07-17T08:34:59.3645354Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/sleef'... 2025-07-17T08:34:59.3645661Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/pybind11'... 2025-07-17T08:34:59.3645971Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/mimalloc'... 2025-07-17T08:34:59.3646282Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto'... 2025-07-17T08:34:59.3646586Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/fmt'... 2025-07-17T08:34:59.3646901Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/cudnn_frontend'... 2025-07-17T08:34:59.3647267Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/VulkanMemoryAllocator'... 2025-07-17T08:34:59.3648160Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/flatbuffers'... 2025-07-17T08:34:59.3648485Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/fbgemm'... 2025-07-17T08:34:59.3648783Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/cutlass'... 2025-07-17T08:34:59.3649089Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/onnx'... 2025-07-17T08:34:59.3649415Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/composable_kernel'... 2025-07-17T08:34:59.3649734Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/aiter'... 2025-07-17T08:34:59.3650220Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/FP16'... 2025-07-17T08:34:59.3650512Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/FXdiv'... 2025-07-17T08:34:59.3650821Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/NNPACK'... 2025-07-17T08:34:59.4641683Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/NVTX'... 2025-07-17T08:34:59.5819552Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/XNNPACK'... 2025-07-17T08:35:05.2196465Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp'... 2025-07-17T08:35:05.2196902Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/nlohmann'... 2025-07-17T08:35:05.2197243Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/protobuf'... 2025-07-17T08:35:05.2459563Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-07-17T08:35:05.2671660Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-07-17T08:35:05.2822797Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-07-17T08:35:05.3238690Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-07-17T08:35:05.4440881Z Submodule path 'third_party/NVTX': checked out '2942f167cc30c5e3a44a2aecd5b0d9c07ff61a07' 2025-07-17T08:35:05.5188580Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-07-17T08:35:06.7685817Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-07-17T08:35:06.9787731Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-07-17T08:35:06.9826732Z Submodule '3rdparty/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/aiter/3rdparty/composable_kernel' 2025-07-17T08:35:06.9867738Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/aiter/3rdparty/composable_kernel'... 2025-07-17T08:35:09.9785745Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-07-17T08:35:10.0199309Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-07-17T08:35:10.5298591Z Submodule path 'third_party/composable_kernel': checked out '434d19f696da62c12b5372b32cbc9ba968588d7e' 2025-07-17T08:35:10.6029532Z Submodule path 'third_party/cpp-httplib': checked out '3af7f2c16147f3fbc6e4d717032daf505dc1652c' 2025-07-17T08:35:10.7420115Z Submodule path 'third_party/cpuinfo': checked out '5e3d2445e6a84d9599bee2bf78edbb4d80865e1d' 2025-07-17T08:35:10.9048487Z Submodule path 'third_party/cudnn_frontend': checked out 'f937055efc6d414d11f4c6577e3977fe74f35fb6' 2025-07-17T08:35:12.0049225Z Submodule path 'third_party/cutlass': checked out 'b995f933179c22d3fe0d871c3a53d11e4681950f' 2025-07-17T08:35:12.2075594Z Submodule path 'third_party/fbgemm': checked out '157e88b750c452bef2ab4653fe9d1eeb151ce4c3' 2025-07-17T08:35:12.2115442Z Submodule 'external/asmjit' (https://github.com/asmjit/asmjit.git) registered for path 'third_party/fbgemm/external/asmjit' 2025-07-17T08:35:12.2121190Z Submodule 'external/composable_kernel' (https://github.com/jwfromm/composable_kernel.git) registered for path 'third_party/fbgemm/external/composable_kernel' 2025-07-17T08:35:12.2127812Z Submodule 'external/cpuinfo' (https://github.com/pytorch/cpuinfo) registered for path 'third_party/fbgemm/external/cpuinfo' 2025-07-17T08:35:12.2133914Z Submodule 'external/cutlass' (https://github.com/jwfromm/cutlass) registered for path 'third_party/fbgemm/external/cutlass' 2025-07-17T08:35:12.2140282Z Submodule 'external/googletest' (https://github.com/google/googletest) registered for path 'third_party/fbgemm/external/googletest' 2025-07-17T08:35:12.2146956Z Submodule 'external/hipify_torch' (https://github.com/ROCmSoftwarePlatform/hipify_torch.git) registered for path 'third_party/fbgemm/external/hipify_torch' 2025-07-17T08:35:12.2153630Z Submodule 'external/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/fbgemm/external/json' 2025-07-17T08:35:12.2201545Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/fbgemm/external/asmjit'... 2025-07-17T08:35:13.3855390Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/fbgemm/external/hipify_torch'... 2025-07-17T08:35:13.3856023Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/fbgemm/external/cpuinfo'... 2025-07-17T08:35:13.3856509Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/fbgemm/external/googletest'... 2025-07-17T08:35:13.4840904Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/fbgemm/external/composable_kernel'... 2025-07-17T08:35:13.5841875Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/fbgemm/external/cutlass'... 2025-07-17T08:35:14.4712862Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/fbgemm/external/json'... 2025-07-17T08:35:18.0526092Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'e5d7c0bd5d9aec44d68830187138149e6a8c4e32' 2025-07-17T08:35:18.4750581Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out '4a61bdd4bd4ed730e078aebc7c0fcf046ff29406' 2025-07-17T08:35:18.6170706Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-07-17T08:35:19.2679001Z From https://github.com/jwfromm/cutlass 2025-07-17T08:35:19.2679345Z * branch 3ed8d2ec4ba35ef5d9d8353826209b6f868f63d3 -> FETCH_HEAD 2025-07-17T08:35:20.1956958Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '3ed8d2ec4ba35ef5d9d8353826209b6f868f63d3' 2025-07-17T08:35:20.2623500Z Submodule path 'third_party/fbgemm/external/googletest': checked out 'f8d7d77c06936315286eb55f8de22cd23c188571' 2025-07-17T08:35:20.2808484Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out 'a4337c69fe0e2552a7b7b0669178926beeed828c' 2025-07-17T08:35:20.4669426Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-07-17T08:35:20.5996070Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-07-17T08:35:20.6032635Z Submodule 'csrc/composable_kernel' (https://github.com/ROCm/composable_kernel.git) registered for path 'third_party/flash-attention/csrc/composable_kernel' 2025-07-17T08:35:20.6038231Z Submodule 'csrc/cutlass' (https://github.com/NVIDIA/cutlass.git) registered for path 'third_party/flash-attention/csrc/cutlass' 2025-07-17T08:35:20.6081110Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/composable_kernel'... 2025-07-17T08:35:23.1323952Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/flash-attention/csrc/cutlass'... 2025-07-17T08:35:23.6171408Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-07-17T08:35:24.5995399Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-07-17T08:35:24.8619672Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-07-17T08:35:24.9113607Z Submodule path 'third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-07-17T08:35:24.9673143Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-07-17T08:35:25.0131899Z Submodule path 'third_party/gloo': checked out 'c7b7b022c124d9643957d9bd55f57ac59fce8fa2' 2025-07-17T08:35:25.0814814Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-07-17T08:35:25.1034405Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-07-17T08:35:25.1058008Z Submodule 'mkl-dnn' (https://github.com/intel/mkl-dnn.git) registered for path 'third_party/ideep/mkl-dnn' 2025-07-17T08:35:25.1092521Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/ideep/mkl-dnn'... 2025-07-17T08:35:35.2203133Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-07-17T08:35:35.2573716Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-07-17T08:35:35.3925712Z Submodule path 'third_party/kineto': checked out '5e7501833f1021ce6f618572d3baf657b6319658' 2025-07-17T08:35:35.3961944Z Submodule 'libkineto/third_party/dynolog' (https://github.com/facebookincubator/dynolog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog' 2025-07-17T08:35:35.3967720Z Submodule 'libkineto/third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/fmt' 2025-07-17T08:35:35.3974166Z Submodule 'libkineto/third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/googletest' 2025-07-17T08:35:35.4021442Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog'... 2025-07-17T08:35:36.3049724Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/fmt'... 2025-07-17T08:35:36.8327438Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/googletest'... 2025-07-17T08:35:36.9531684Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out '7d04a0053a845370ae06ce317a22a48e9edcc74e' 2025-07-17T08:35:36.9565688Z Submodule 'third_party/DCGM' (https://github.com/NVIDIA/DCGM.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-07-17T08:35:36.9571692Z Submodule 'third_party/cpr' (https://github.com/libcpr/cpr.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-07-17T08:35:36.9577966Z Submodule 'third_party/fmt' (https://github.com/fmtlib/fmt.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-07-17T08:35:36.9585336Z Submodule 'third_party/gflags' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-07-17T08:35:36.9592069Z Submodule 'third_party/glog' (https://github.com/google/glog.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-07-17T08:35:36.9599385Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-07-17T08:35:36.9606449Z Submodule 'third_party/json' (https://github.com/nlohmann/json.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-07-17T08:35:36.9613849Z Submodule 'third_party/pfs' (https://github.com/dtrugman/pfs.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-07-17T08:35:36.9661158Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM'... 2025-07-17T08:35:38.2469811Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/pfs'... 2025-07-17T08:35:38.2470471Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags'... 2025-07-17T08:35:38.2471587Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/cpr'... 2025-07-17T08:35:38.2472143Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/glog'... 2025-07-17T08:35:38.2472700Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/googletest'... 2025-07-17T08:35:38.2920263Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/fmt'... 2025-07-17T08:35:38.3922230Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/json'... 2025-07-17T08:35:42.7289122Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-07-17T08:35:42.7617739Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-07-17T08:35:42.8182069Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-07-17T08:35:42.8408584Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-07-17T08:35:42.8435637Z Submodule 'doc' (https://github.com/gflags/gflags.git) registered for path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-07-17T08:35:42.8475793Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc'... 2025-07-17T08:35:43.3729724Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-07-17T08:35:43.4030862Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-07-17T08:35:43.4659839Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '58d77fa8070e8cec2dc1ed015d66b454c8d78850' 2025-07-17T08:35:43.6470125Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-07-17T08:35:43.6754365Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-07-17T08:35:43.7299203Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '0041a40c1350ba702d475b9c4ad62da77caea164' 2025-07-17T08:35:43.8100104Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '7aca84427f224eeed3144123d5230d5871e93347' 2025-07-17T08:35:43.8850690Z Submodule path 'third_party/kleidiai': checked out 'cca02c2f69dd18e1f12647c1c0bdc8cf90e680c7' 2025-07-17T08:35:43.9512062Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-07-17T08:35:44.1592694Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-07-17T08:35:45.1489437Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-07-17T08:35:45.1562355Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/onnx/third_party/pybind11' 2025-07-17T08:35:45.1608151Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/onnx/third_party/pybind11'... 2025-07-17T08:35:46.2682703Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-07-17T08:35:46.4051970Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-07-17T08:35:46.4089876Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark) registered for path 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-07-17T08:35:46.4096211Z Submodule 'third_party/googletest' (https://github.com/google/googletest) registered for path 'third_party/opentelemetry-cpp/third_party/googletest' 2025-07-17T08:35:46.4101601Z Submodule 'third_party/ms-gsl' (https://github.com/microsoft/GSL) registered for path 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-07-17T08:35:46.4108356Z Submodule 'third_party/nlohmann-json' (https://github.com/nlohmann/json) registered for path 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-07-17T08:35:46.4115581Z Submodule 'third_party/opentelemetry-proto' (https://github.com/open-telemetry/opentelemetry-proto) registered for path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-07-17T08:35:46.4122644Z Submodule 'third_party/opentracing-cpp' (https://github.com/opentracing/opentracing-cpp.git) registered for path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-07-17T08:35:46.4129617Z Submodule 'third_party/prometheus-cpp' (https://github.com/jupp0r/prometheus-cpp) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-07-17T08:35:46.4136479Z Submodule 'tools/vcpkg' (https://github.com/Microsoft/vcpkg) registered for path 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-07-17T08:35:46.4184329Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/benchmark'... 2025-07-17T08:35:47.0466352Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentracing-cpp'... 2025-07-17T08:35:47.0466994Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/opentelemetry-proto'... 2025-07-17T08:35:47.0467512Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/ms-gsl'... 2025-07-17T08:35:47.0467996Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp'... 2025-07-17T08:35:47.1468038Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/googletest'... 2025-07-17T08:35:47.5804347Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/nlohmann-json'... 2025-07-17T08:35:52.1086725Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/tools/vcpkg'... 2025-07-17T08:35:52.6165355Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-07-17T08:35:52.6760883Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-07-17T08:35:52.6998862Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-07-17T08:35:52.8894030Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-07-17T08:35:52.9126992Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-07-17T08:35:52.9378395Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-07-17T08:35:52.9820120Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-07-17T08:35:52.9866870Z Submodule 'civetweb' (https://github.com/civetweb/civetweb.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-07-17T08:35:52.9876896Z Submodule 'googletest' (https://github.com/google/googletest.git) registered for path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-07-17T08:35:52.9923966Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb'... 2025-07-17T08:35:55.5829604Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest'... 2025-07-17T08:35:55.9350331Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-07-17T08:35:56.0033810Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-07-17T08:35:57.1167745Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-07-17T08:35:57.1612702Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-07-17T08:35:57.7015062Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-07-17T08:35:57.7057774Z Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'third_party/protobuf/third_party/benchmark' 2025-07-17T08:35:57.7062801Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/protobuf/third_party/googletest' 2025-07-17T08:35:57.7104150Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/protobuf/third_party/benchmark'... 2025-07-17T08:35:58.4531559Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/protobuf/third_party/googletest'... 2025-07-17T08:35:58.7561736Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-07-17T08:35:58.8472279Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-07-17T08:35:58.8630932Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-07-17T08:35:58.8827367Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-07-17T08:35:58.9425970Z Submodule path 'third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-07-17T08:35:58.9892973Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-07-17T08:35:59.0587572Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-07-17T08:35:59.1060191Z Submodule path 'third_party/tensorpipe': checked out '52791a2fd214b2a9dc5759d36725909c1daa7f2e' 2025-07-17T08:35:59.1091863Z Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'third_party/tensorpipe/third_party/googletest' 2025-07-17T08:35:59.1097053Z Submodule 'third_party/libnop' (https://github.com/google/libnop.git) registered for path 'third_party/tensorpipe/third_party/libnop' 2025-07-17T08:35:59.1103043Z Submodule 'third_party/libuv' (https://github.com/libuv/libuv.git) registered for path 'third_party/tensorpipe/third_party/libuv' 2025-07-17T08:35:59.1109623Z Submodule 'third_party/pybind11' (https://github.com/pybind/pybind11.git) registered for path 'third_party/tensorpipe/third_party/pybind11' 2025-07-17T08:35:59.1149377Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/googletest'... 2025-07-17T08:36:00.2384507Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libnop'... 2025-07-17T08:36:00.6512672Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/libuv'... 2025-07-17T08:36:01.1994496Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11'... 2025-07-17T08:36:01.2736692Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-07-17T08:36:01.3004734Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-07-17T08:36:01.3992531Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '1dff88e5161cba5c59276d2070d2e304e4dcb242' 2025-07-17T08:36:01.4430764Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-07-17T08:36:01.4457368Z Submodule 'tools/clang' (https://github.com/wjakob/clang-cindex-python3) registered for path 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-07-17T08:36:01.4494203Z Cloning into '/home/runner/_work/pytorch/pytorch/third_party/tensorpipe/third_party/pybind11/tools/clang'... 2025-07-17T08:36:01.8461490Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-07-17T08:36:01.8518979Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-07-17T08:36:01.8826118Z Entering 'android/libs/fbjni' 2025-07-17T08:36:01.8877157Z Entering 'third_party/FP16' 2025-07-17T08:36:01.8926607Z Entering 'third_party/FXdiv' 2025-07-17T08:36:01.8973097Z Entering 'third_party/NNPACK' 2025-07-17T08:36:01.9019612Z Entering 'third_party/NVTX' 2025-07-17T08:36:01.9068319Z Entering 'third_party/VulkanMemoryAllocator' 2025-07-17T08:36:01.9117050Z Entering 'third_party/XNNPACK' 2025-07-17T08:36:01.9180694Z Entering 'third_party/aiter' 2025-07-17T08:36:01.9229754Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-07-17T08:36:01.9286784Z Entering 'third_party/benchmark' 2025-07-17T08:36:01.9334388Z Entering 'third_party/composable_kernel' 2025-07-17T08:36:01.9389311Z Entering 'third_party/cpp-httplib' 2025-07-17T08:36:01.9436045Z Entering 'third_party/cpuinfo' 2025-07-17T08:36:01.9483958Z Entering 'third_party/cudnn_frontend' 2025-07-17T08:36:01.9530874Z Entering 'third_party/cutlass' 2025-07-17T08:36:01.9586872Z Entering 'third_party/fbgemm' 2025-07-17T08:36:01.9637270Z Entering 'third_party/fbgemm/external/asmjit' 2025-07-17T08:36:01.9682257Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-07-17T08:36:01.9737926Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-07-17T08:36:01.9783628Z Entering 'third_party/fbgemm/external/cutlass' 2025-07-17T08:36:01.9844956Z Entering 'third_party/fbgemm/external/googletest' 2025-07-17T08:36:01.9892305Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-07-17T08:36:01.9938014Z Entering 'third_party/fbgemm/external/json' 2025-07-17T08:36:01.9990750Z Entering 'third_party/flash-attention' 2025-07-17T08:36:02.0040858Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-07-17T08:36:02.0093176Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-07-17T08:36:02.0148432Z Entering 'third_party/flatbuffers' 2025-07-17T08:36:02.0199151Z Entering 'third_party/fmt' 2025-07-17T08:36:02.0247619Z Entering 'third_party/gemmlowp/gemmlowp' 2025-07-17T08:36:02.0294331Z Entering 'third_party/gloo' 2025-07-17T08:36:02.0341378Z Entering 'third_party/googletest' 2025-07-17T08:36:02.0390855Z Entering 'third_party/ideep' 2025-07-17T08:36:02.0436413Z Entering 'third_party/ideep/mkl-dnn' 2025-07-17T08:36:02.0490179Z Entering 'third_party/ittapi' 2025-07-17T08:36:02.0535500Z Entering 'third_party/kineto' 2025-07-17T08:36:02.0589978Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-07-17T08:36:02.0635547Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-07-17T08:36:02.0683098Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-07-17T08:36:02.0732469Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-07-17T08:36:02.0778698Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-07-17T08:36:02.0832300Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-07-17T08:36:02.0883711Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-07-17T08:36:02.0930831Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-07-17T08:36:02.0978588Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-07-17T08:36:02.1026717Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-07-17T08:36:02.1079055Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-07-17T08:36:02.1127520Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-07-17T08:36:02.1178300Z Entering 'third_party/kleidiai' 2025-07-17T08:36:02.1226418Z Entering 'third_party/mimalloc' 2025-07-17T08:36:02.1274799Z Entering 'third_party/nlohmann' 2025-07-17T08:36:02.1322033Z Entering 'third_party/onnx' 2025-07-17T08:36:02.1385862Z Entering 'third_party/onnx/third_party/pybind11' 2025-07-17T08:36:02.1436825Z Entering 'third_party/opentelemetry-cpp' 2025-07-17T08:36:02.1485029Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-07-17T08:36:02.1529149Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-07-17T08:36:02.1574205Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-07-17T08:36:02.1622573Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-07-17T08:36:02.1671993Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-07-17T08:36:02.1718556Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-07-17T08:36:02.1764764Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-07-17T08:36:02.1807250Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-07-17T08:36:02.1853845Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-07-17T08:36:02.1901354Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-07-17T08:36:02.1973817Z Entering 'third_party/pocketfft' 2025-07-17T08:36:02.2022193Z Entering 'third_party/protobuf' 2025-07-17T08:36:02.2074333Z Entering 'third_party/protobuf/third_party/benchmark' 2025-07-17T08:36:02.2119050Z Entering 'third_party/protobuf/third_party/googletest' 2025-07-17T08:36:02.2168282Z Entering 'third_party/psimd' 2025-07-17T08:36:02.2216920Z Entering 'third_party/pthreadpool' 2025-07-17T08:36:02.2263019Z Entering 'third_party/pybind11' 2025-07-17T08:36:02.2310469Z Entering 'third_party/python-peachpy' 2025-07-17T08:36:02.2358446Z Entering 'third_party/sleef' 2025-07-17T08:36:02.2406041Z Entering 'third_party/tensorpipe' 2025-07-17T08:36:02.2454616Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-07-17T08:36:02.2500197Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-07-17T08:36:02.2548084Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-07-17T08:36:02.2594548Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-07-17T08:36:02.2639265Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-07-17T08:36:02.2707621Z ##[endgroup] 2025-07-17T08:36:02.2707915Z ##[group]Persisting credentials for submodules 2025-07-17T08:36:02.2723678Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-07-17T08:36:02.3024478Z Entering 'android/libs/fbjni' 2025-07-17T08:36:02.3078005Z Entering 'third_party/FP16' 2025-07-17T08:36:02.3130671Z Entering 'third_party/FXdiv' 2025-07-17T08:36:02.3182504Z Entering 'third_party/NNPACK' 2025-07-17T08:36:02.3236208Z Entering 'third_party/NVTX' 2025-07-17T08:36:02.3287908Z Entering 'third_party/VulkanMemoryAllocator' 2025-07-17T08:36:02.3341224Z Entering 'third_party/XNNPACK' 2025-07-17T08:36:02.3407433Z Entering 'third_party/aiter' 2025-07-17T08:36:02.3461357Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-07-17T08:36:02.3523967Z Entering 'third_party/benchmark' 2025-07-17T08:36:02.3574973Z Entering 'third_party/composable_kernel' 2025-07-17T08:36:02.3632808Z Entering 'third_party/cpp-httplib' 2025-07-17T08:36:02.3685723Z Entering 'third_party/cpuinfo' 2025-07-17T08:36:02.3738583Z Entering 'third_party/cudnn_frontend' 2025-07-17T08:36:02.3789812Z Entering 'third_party/cutlass' 2025-07-17T08:36:02.3849489Z Entering 'third_party/fbgemm' 2025-07-17T08:36:02.3902877Z Entering 'third_party/fbgemm/external/asmjit' 2025-07-17T08:36:02.3954004Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-07-17T08:36:02.4008873Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-07-17T08:36:02.4061655Z Entering 'third_party/fbgemm/external/cutlass' 2025-07-17T08:36:02.4123945Z Entering 'third_party/fbgemm/external/googletest' 2025-07-17T08:36:02.4174006Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-07-17T08:36:02.4223031Z Entering 'third_party/fbgemm/external/json' 2025-07-17T08:36:02.4277681Z Entering 'third_party/flash-attention' 2025-07-17T08:36:02.4334846Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-07-17T08:36:02.4390446Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-07-17T08:36:02.4452483Z Entering 'third_party/flatbuffers' 2025-07-17T08:36:02.4506883Z Entering 'third_party/fmt' 2025-07-17T08:36:02.4560588Z Entering 'third_party/gemmlowp/gemmlowp' 2025-07-17T08:36:02.4611476Z Entering 'third_party/gloo' 2025-07-17T08:36:02.4663122Z Entering 'third_party/googletest' 2025-07-17T08:36:02.4715324Z Entering 'third_party/ideep' 2025-07-17T08:36:02.4766528Z Entering 'third_party/ideep/mkl-dnn' 2025-07-17T08:36:02.4823723Z Entering 'third_party/ittapi' 2025-07-17T08:36:02.4875052Z Entering 'third_party/kineto' 2025-07-17T08:36:02.4928541Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-07-17T08:36:02.4978147Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-07-17T08:36:02.5033498Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-07-17T08:36:02.5086716Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-07-17T08:36:02.5136857Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-07-17T08:36:02.5191314Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-07-17T08:36:02.5246300Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-07-17T08:36:02.5299034Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-07-17T08:36:02.5350902Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-07-17T08:36:02.5405240Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-07-17T08:36:02.5460916Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-07-17T08:36:02.5512119Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-07-17T08:36:02.5566657Z Entering 'third_party/kleidiai' 2025-07-17T08:36:02.5618174Z Entering 'third_party/mimalloc' 2025-07-17T08:36:02.5669094Z Entering 'third_party/nlohmann' 2025-07-17T08:36:02.5721700Z Entering 'third_party/onnx' 2025-07-17T08:36:02.5788667Z Entering 'third_party/onnx/third_party/pybind11' 2025-07-17T08:36:02.5847468Z Entering 'third_party/opentelemetry-cpp' 2025-07-17T08:36:02.5900491Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-07-17T08:36:02.5952157Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-07-17T08:36:02.6003317Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-07-17T08:36:02.6054122Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-07-17T08:36:02.6106468Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-07-17T08:36:02.6155747Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-07-17T08:36:02.6206849Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-07-17T08:36:02.6256615Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-07-17T08:36:02.6307912Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-07-17T08:36:02.6361904Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-07-17T08:36:02.6428848Z Entering 'third_party/pocketfft' 2025-07-17T08:36:02.6478212Z Entering 'third_party/protobuf' 2025-07-17T08:36:02.6533591Z Entering 'third_party/protobuf/third_party/benchmark' 2025-07-17T08:36:02.6583081Z Entering 'third_party/protobuf/third_party/googletest' 2025-07-17T08:36:02.6639201Z Entering 'third_party/psimd' 2025-07-17T08:36:02.6692227Z Entering 'third_party/pthreadpool' 2025-07-17T08:36:02.6744515Z Entering 'third_party/pybind11' 2025-07-17T08:36:02.6796262Z Entering 'third_party/python-peachpy' 2025-07-17T08:36:02.6848372Z Entering 'third_party/sleef' 2025-07-17T08:36:02.6898184Z Entering 'third_party/tensorpipe' 2025-07-17T08:36:02.6950792Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-07-17T08:36:02.7001031Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-07-17T08:36:02.7050802Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-07-17T08:36:02.7101033Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-07-17T08:36:02.7150941Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-07-17T08:36:02.7223458Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-07-17T08:36:02.7519465Z Entering 'android/libs/fbjni' 2025-07-17T08:36:02.7567723Z file:/home/runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-07-17T08:36:02.7591797Z Entering 'third_party/FP16' 2025-07-17T08:36:02.7640849Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-07-17T08:36:02.7664091Z Entering 'third_party/FXdiv' 2025-07-17T08:36:02.7710929Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-07-17T08:36:02.7734490Z Entering 'third_party/NNPACK' 2025-07-17T08:36:02.7781130Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-07-17T08:36:02.7804947Z Entering 'third_party/NVTX' 2025-07-17T08:36:02.7850950Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-07-17T08:36:02.7875960Z Entering 'third_party/VulkanMemoryAllocator' 2025-07-17T08:36:02.7920975Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-07-17T08:36:02.7943715Z Entering 'third_party/XNNPACK' 2025-07-17T08:36:02.7990388Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-07-17T08:36:02.8029521Z Entering 'third_party/aiter' 2025-07-17T08:36:02.8075463Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-07-17T08:36:02.8099362Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-07-17T08:36:02.8143459Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-07-17T08:36:02.8177590Z Entering 'third_party/benchmark' 2025-07-17T08:36:02.8225766Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-07-17T08:36:02.8249504Z Entering 'third_party/composable_kernel' 2025-07-17T08:36:02.8296531Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-07-17T08:36:02.8327551Z Entering 'third_party/cpp-httplib' 2025-07-17T08:36:02.8373689Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-07-17T08:36:02.8399907Z Entering 'third_party/cpuinfo' 2025-07-17T08:36:02.8446049Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-07-17T08:36:02.8470387Z Entering 'third_party/cudnn_frontend' 2025-07-17T08:36:02.8517011Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-07-17T08:36:02.8539713Z Entering 'third_party/cutlass' 2025-07-17T08:36:02.8587793Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-07-17T08:36:02.8621127Z Entering 'third_party/fbgemm' 2025-07-17T08:36:02.8667433Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-07-17T08:36:02.8694850Z Entering 'third_party/fbgemm/external/asmjit' 2025-07-17T08:36:02.8740953Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-07-17T08:36:02.8765832Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-07-17T08:36:02.8815662Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-07-17T08:36:02.8845231Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-07-17T08:36:02.8889420Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-07-17T08:36:02.8913492Z Entering 'third_party/fbgemm/external/cutlass' 2025-07-17T08:36:02.8962471Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-07-17T08:36:02.8993325Z Entering 'third_party/fbgemm/external/googletest' 2025-07-17T08:36:02.9040061Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-07-17T08:36:02.9063087Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-07-17T08:36:02.9109265Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-07-17T08:36:02.9132756Z Entering 'third_party/fbgemm/external/json' 2025-07-17T08:36:02.9180969Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-07-17T08:36:02.9208112Z Entering 'third_party/flash-attention' 2025-07-17T08:36:02.9255954Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-07-17T08:36:02.9280639Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-07-17T08:36:02.9329131Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-07-17T08:36:02.9359734Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-07-17T08:36:02.9406235Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-07-17T08:36:02.9439668Z Entering 'third_party/flatbuffers' 2025-07-17T08:36:02.9486616Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-07-17T08:36:02.9513097Z Entering 'third_party/fmt' 2025-07-17T08:36:02.9561242Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-07-17T08:36:02.9585290Z Entering 'third_party/gemmlowp/gemmlowp' 2025-07-17T08:36:02.9633376Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-07-17T08:36:02.9656343Z Entering 'third_party/gloo' 2025-07-17T08:36:02.9703957Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-07-17T08:36:02.9727656Z Entering 'third_party/googletest' 2025-07-17T08:36:02.9772329Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-07-17T08:36:02.9796473Z Entering 'third_party/ideep' 2025-07-17T08:36:02.9843062Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-07-17T08:36:02.9866153Z Entering 'third_party/ideep/mkl-dnn' 2025-07-17T08:36:02.9913475Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-07-17T08:36:02.9946323Z Entering 'third_party/ittapi' 2025-07-17T08:36:02.9992913Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-07-17T08:36:03.0017282Z Entering 'third_party/kineto' 2025-07-17T08:36:03.0065694Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-07-17T08:36:03.0089481Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-07-17T08:36:03.0135597Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-07-17T08:36:03.0161717Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-07-17T08:36:03.0206790Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-07-17T08:36:03.0232879Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-07-17T08:36:03.0281193Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-07-17T08:36:03.0305734Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-07-17T08:36:03.0355557Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-07-17T08:36:03.0379247Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-07-17T08:36:03.0427730Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-07-17T08:36:03.0451022Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-07-17T08:36:03.0500000Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-07-17T08:36:03.0528003Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-07-17T08:36:03.0575278Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-07-17T08:36:03.0599520Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-07-17T08:36:03.0647255Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-07-17T08:36:03.0670096Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-07-17T08:36:03.0717455Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-07-17T08:36:03.0743726Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-07-17T08:36:03.0790246Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-07-17T08:36:03.0818484Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-07-17T08:36:03.0864201Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-07-17T08:36:03.0889710Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-07-17T08:36:03.0936256Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-07-17T08:36:03.0964129Z Entering 'third_party/kleidiai' 2025-07-17T08:36:03.1010646Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-07-17T08:36:03.1036945Z Entering 'third_party/mimalloc' 2025-07-17T08:36:03.1081920Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-07-17T08:36:03.1105922Z Entering 'third_party/nlohmann' 2025-07-17T08:36:03.1152432Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-07-17T08:36:03.1176916Z Entering 'third_party/onnx' 2025-07-17T08:36:03.1224461Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-07-17T08:36:03.1263246Z Entering 'third_party/onnx/third_party/pybind11' 2025-07-17T08:36:03.1310536Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-07-17T08:36:03.1338780Z Entering 'third_party/opentelemetry-cpp' 2025-07-17T08:36:03.1385433Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-07-17T08:36:03.1411686Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-07-17T08:36:03.1457653Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-07-17T08:36:03.1482676Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-07-17T08:36:03.1529569Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-07-17T08:36:03.1551563Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-07-17T08:36:03.1595542Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-07-17T08:36:03.1619669Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-07-17T08:36:03.1664607Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-07-17T08:36:03.1688163Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-07-17T08:36:03.1734750Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-07-17T08:36:03.1757221Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-07-17T08:36:03.1803567Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-07-17T08:36:03.1827353Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-07-17T08:36:03.1870751Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-07-17T08:36:03.1892770Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-07-17T08:36:03.1938665Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-07-17T08:36:03.1963783Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-07-17T08:36:03.2006892Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-07-17T08:36:03.2034791Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-07-17T08:36:03.2078768Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-07-17T08:36:03.2121109Z Entering 'third_party/pocketfft' 2025-07-17T08:36:03.2165232Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-07-17T08:36:03.2187724Z Entering 'third_party/protobuf' 2025-07-17T08:36:03.2233354Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-07-17T08:36:03.2258300Z Entering 'third_party/protobuf/third_party/benchmark' 2025-07-17T08:36:03.2303713Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-07-17T08:36:03.2329237Z Entering 'third_party/protobuf/third_party/googletest' 2025-07-17T08:36:03.2373355Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-07-17T08:36:03.2400713Z Entering 'third_party/psimd' 2025-07-17T08:36:03.2450300Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-07-17T08:36:03.2473813Z Entering 'third_party/pthreadpool' 2025-07-17T08:36:03.2520617Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-07-17T08:36:03.2544343Z Entering 'third_party/pybind11' 2025-07-17T08:36:03.2589990Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-07-17T08:36:03.2615199Z Entering 'third_party/python-peachpy' 2025-07-17T08:36:03.2662471Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-07-17T08:36:03.2686595Z Entering 'third_party/sleef' 2025-07-17T08:36:03.2731002Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-07-17T08:36:03.2755024Z Entering 'third_party/tensorpipe' 2025-07-17T08:36:03.2801216Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-07-17T08:36:03.2822301Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-07-17T08:36:03.2869090Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-07-17T08:36:03.2892918Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-07-17T08:36:03.2939887Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-07-17T08:36:03.2962634Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-07-17T08:36:03.3006503Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-07-17T08:36:03.3029861Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-07-17T08:36:03.3075204Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-07-17T08:36:03.3096477Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-07-17T08:36:03.3142099Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-07-17T08:36:03.3427602Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-07-17T08:36:03.3732190Z Entering 'android/libs/fbjni' 2025-07-17T08:36:03.3783967Z Entering 'third_party/FP16' 2025-07-17T08:36:03.3833963Z Entering 'third_party/FXdiv' 2025-07-17T08:36:03.3881566Z Entering 'third_party/NNPACK' 2025-07-17T08:36:03.3928147Z Entering 'third_party/NVTX' 2025-07-17T08:36:03.3975961Z Entering 'third_party/VulkanMemoryAllocator' 2025-07-17T08:36:03.4025059Z Entering 'third_party/XNNPACK' 2025-07-17T08:36:03.4089311Z Entering 'third_party/aiter' 2025-07-17T08:36:03.4139122Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-07-17T08:36:03.4195590Z Entering 'third_party/benchmark' 2025-07-17T08:36:03.4243409Z Entering 'third_party/composable_kernel' 2025-07-17T08:36:03.4297792Z Entering 'third_party/cpp-httplib' 2025-07-17T08:36:03.4350095Z Entering 'third_party/cpuinfo' 2025-07-17T08:36:03.4398798Z Entering 'third_party/cudnn_frontend' 2025-07-17T08:36:03.4445600Z Entering 'third_party/cutlass' 2025-07-17T08:36:03.4501043Z Entering 'third_party/fbgemm' 2025-07-17T08:36:03.4549467Z Entering 'third_party/fbgemm/external/asmjit' 2025-07-17T08:36:03.4596059Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-07-17T08:36:03.4655296Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-07-17T08:36:03.4702132Z Entering 'third_party/fbgemm/external/cutlass' 2025-07-17T08:36:03.4756367Z Entering 'third_party/fbgemm/external/googletest' 2025-07-17T08:36:03.4800963Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-07-17T08:36:03.4847337Z Entering 'third_party/fbgemm/external/json' 2025-07-17T08:36:03.4897560Z Entering 'third_party/flash-attention' 2025-07-17T08:36:03.4945785Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-07-17T08:36:03.5000230Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-07-17T08:36:03.5059336Z Entering 'third_party/flatbuffers' 2025-07-17T08:36:03.5110583Z Entering 'third_party/fmt' 2025-07-17T08:36:03.5159875Z Entering 'third_party/gemmlowp/gemmlowp' 2025-07-17T08:36:03.5208878Z Entering 'third_party/gloo' 2025-07-17T08:36:03.5253600Z Entering 'third_party/googletest' 2025-07-17T08:36:03.5300139Z Entering 'third_party/ideep' 2025-07-17T08:36:03.5351481Z Entering 'third_party/ideep/mkl-dnn' 2025-07-17T08:36:03.5405677Z Entering 'third_party/ittapi' 2025-07-17T08:36:03.5453166Z Entering 'third_party/kineto' 2025-07-17T08:36:03.5502644Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-07-17T08:36:03.5548438Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-07-17T08:36:03.5600969Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-07-17T08:36:03.5646520Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-07-17T08:36:03.5692110Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-07-17T08:36:03.5736937Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-07-17T08:36:03.5791861Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-07-17T08:36:03.5838728Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-07-17T08:36:03.5886222Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-07-17T08:36:03.5934388Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-07-17T08:36:03.5984280Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-07-17T08:36:03.6030514Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-07-17T08:36:03.6080917Z Entering 'third_party/kleidiai' 2025-07-17T08:36:03.6129118Z Entering 'third_party/mimalloc' 2025-07-17T08:36:03.6174151Z Entering 'third_party/nlohmann' 2025-07-17T08:36:03.6223879Z Entering 'third_party/onnx' 2025-07-17T08:36:03.6289994Z Entering 'third_party/onnx/third_party/pybind11' 2025-07-17T08:36:03.6340199Z Entering 'third_party/opentelemetry-cpp' 2025-07-17T08:36:03.6390513Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-07-17T08:36:03.6437487Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-07-17T08:36:03.6482276Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-07-17T08:36:03.6527594Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-07-17T08:36:03.6572835Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-07-17T08:36:03.6621550Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-07-17T08:36:03.6666505Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-07-17T08:36:03.6711532Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-07-17T08:36:03.6760547Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-07-17T08:36:03.6810639Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-07-17T08:36:03.6874418Z Entering 'third_party/pocketfft' 2025-07-17T08:36:03.6924317Z Entering 'third_party/protobuf' 2025-07-17T08:36:03.6973757Z Entering 'third_party/protobuf/third_party/benchmark' 2025-07-17T08:36:03.7019882Z Entering 'third_party/protobuf/third_party/googletest' 2025-07-17T08:36:03.7071740Z Entering 'third_party/psimd' 2025-07-17T08:36:03.7120191Z Entering 'third_party/pthreadpool' 2025-07-17T08:36:03.7166168Z Entering 'third_party/pybind11' 2025-07-17T08:36:03.7212111Z Entering 'third_party/python-peachpy' 2025-07-17T08:36:03.7258378Z Entering 'third_party/sleef' 2025-07-17T08:36:03.7309264Z Entering 'third_party/tensorpipe' 2025-07-17T08:36:03.7360922Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-07-17T08:36:03.7405544Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-07-17T08:36:03.7449927Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-07-17T08:36:03.7493742Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-07-17T08:36:03.7541150Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-07-17T08:36:03.7609148Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-07-17T08:36:03.7911970Z Entering 'android/libs/fbjni' 2025-07-17T08:36:03.7961826Z Entering 'third_party/FP16' 2025-07-17T08:36:03.8009381Z Entering 'third_party/FXdiv' 2025-07-17T08:36:03.8057483Z Entering 'third_party/NNPACK' 2025-07-17T08:36:03.8106244Z Entering 'third_party/NVTX' 2025-07-17T08:36:03.8154398Z Entering 'third_party/VulkanMemoryAllocator' 2025-07-17T08:36:03.8201051Z Entering 'third_party/XNNPACK' 2025-07-17T08:36:03.8264177Z Entering 'third_party/aiter' 2025-07-17T08:36:03.8324748Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-07-17T08:36:03.8384898Z Entering 'third_party/benchmark' 2025-07-17T08:36:03.8448689Z Entering 'third_party/composable_kernel' 2025-07-17T08:36:03.8510426Z Entering 'third_party/cpp-httplib' 2025-07-17T08:36:03.8571755Z Entering 'third_party/cpuinfo' 2025-07-17T08:36:03.8628817Z Entering 'third_party/cudnn_frontend' 2025-07-17T08:36:03.8678309Z Entering 'third_party/cutlass' 2025-07-17T08:36:03.8747640Z Entering 'third_party/fbgemm' 2025-07-17T08:36:03.8804220Z Entering 'third_party/fbgemm/external/asmjit' 2025-07-17T08:36:03.8864353Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-07-17T08:36:03.8923333Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-07-17T08:36:03.8975391Z Entering 'third_party/fbgemm/external/cutlass' 2025-07-17T08:36:03.9027985Z Entering 'third_party/fbgemm/external/googletest' 2025-07-17T08:36:03.9071179Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-07-17T08:36:03.9111769Z Entering 'third_party/fbgemm/external/json' 2025-07-17T08:36:03.9158826Z Entering 'third_party/flash-attention' 2025-07-17T08:36:03.9206581Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-07-17T08:36:03.9256224Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-07-17T08:36:03.9310462Z Entering 'third_party/flatbuffers' 2025-07-17T08:36:03.9356938Z Entering 'third_party/fmt' 2025-07-17T08:36:03.9402478Z Entering 'third_party/gemmlowp/gemmlowp' 2025-07-17T08:36:03.9451763Z Entering 'third_party/gloo' 2025-07-17T08:36:03.9498016Z Entering 'third_party/googletest' 2025-07-17T08:36:03.9542674Z Entering 'third_party/ideep' 2025-07-17T08:36:03.9586183Z Entering 'third_party/ideep/mkl-dnn' 2025-07-17T08:36:03.9640940Z Entering 'third_party/ittapi' 2025-07-17T08:36:03.9692703Z Entering 'third_party/kineto' 2025-07-17T08:36:03.9742205Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-07-17T08:36:03.9791642Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-07-17T08:36:03.9845098Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-07-17T08:36:03.9898267Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-07-17T08:36:03.9946409Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-07-17T08:36:03.9998915Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-07-17T08:36:04.0050493Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-07-17T08:36:04.0100646Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-07-17T08:36:04.0148042Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-07-17T08:36:04.0196329Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-07-17T08:36:04.0248177Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-07-17T08:36:04.0295536Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-07-17T08:36:04.0351763Z Entering 'third_party/kleidiai' 2025-07-17T08:36:04.0406670Z Entering 'third_party/mimalloc' 2025-07-17T08:36:04.0459566Z Entering 'third_party/nlohmann' 2025-07-17T08:36:04.0512981Z Entering 'third_party/onnx' 2025-07-17T08:36:04.0576984Z Entering 'third_party/onnx/third_party/pybind11' 2025-07-17T08:36:04.0633071Z Entering 'third_party/opentelemetry-cpp' 2025-07-17T08:36:04.0686405Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-07-17T08:36:04.0734271Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-07-17T08:36:04.0781440Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-07-17T08:36:04.0828581Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-07-17T08:36:04.0877101Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-07-17T08:36:04.0924039Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-07-17T08:36:04.0973330Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-07-17T08:36:04.1019485Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-07-17T08:36:04.1068561Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-07-17T08:36:04.1122638Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-07-17T08:36:04.1192602Z Entering 'third_party/pocketfft' 2025-07-17T08:36:04.1245980Z Entering 'third_party/protobuf' 2025-07-17T08:36:04.1299294Z Entering 'third_party/protobuf/third_party/benchmark' 2025-07-17T08:36:04.1349334Z Entering 'third_party/protobuf/third_party/googletest' 2025-07-17T08:36:04.1402758Z Entering 'third_party/psimd' 2025-07-17T08:36:04.1455590Z Entering 'third_party/pthreadpool' 2025-07-17T08:36:04.1507239Z Entering 'third_party/pybind11' 2025-07-17T08:36:04.1559342Z Entering 'third_party/python-peachpy' 2025-07-17T08:36:04.1610812Z Entering 'third_party/sleef' 2025-07-17T08:36:04.1662080Z Entering 'third_party/tensorpipe' 2025-07-17T08:36:04.1715403Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-07-17T08:36:04.1761558Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-07-17T08:36:04.1809408Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-07-17T08:36:04.1854627Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-07-17T08:36:04.1903447Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-07-17T08:36:04.1982242Z ##[endgroup] 2025-07-17T08:36:04.2030452Z [command]/usr/bin/git log -1 --format=%H 2025-07-17T08:36:04.2061836Z a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:36:04.2235918Z Prepare all required actions 2025-07-17T08:36:04.2236283Z Getting action download info 2025-07-17T08:36:04.4260573Z ##[group]Run ./.github/actions/setup-rocm 2025-07-17T08:36:04.4260795Z env: 2025-07-17T08:36:04.4260935Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:04.4261102Z ##[endgroup] 2025-07-17T08:36:04.4282864Z ##[group]Run dpkg -l | grep -E " rocm" 2025-07-17T08:36:04.4283080Z dpkg -l | grep -E " rocm" 2025-07-17T08:36:04.4295312Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:04.4295539Z env: 2025-07-17T08:36:04.4295678Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:04.4295840Z ##[endgroup] 2025-07-17T08:36:04.4394876Z ii rocm-cmake 0.14.0.60401-83~22.04 amd64 rocm-cmake built using CMake 2025-07-17T08:36:04.4395248Z ii rocm-core 6.4.1.60401-83~22.04 amd64 ROCm Runtime software stack 2025-07-17T08:36:04.4395661Z ii rocm-dbgapi 0.77.2.60401-83~22.04 amd64 Library to provide AMD GPU debugger API 2025-07-17T08:36:04.4396087Z ii rocm-debug-agent 2.0.4.60401-83~22.04 amd64 Radeon Open Compute Debug Agent (ROCdebug-agent) 2025-07-17T08:36:04.4396494Z ii rocm-dev 6.4.1.60401-83~22.04 amd64 Radeon Open Compute (ROCm) Runtime software stack 2025-07-17T08:36:04.4396877Z ii rocm-device-libs 1.0.0.60401-83~22.04 amd64 Radeon Open Compute - device libraries 2025-07-17T08:36:04.4397217Z ii rocm-gdb 15.2.60401-83~22.04 amd64 ROCgdb 2025-07-17T08:36:04.4397543Z ii rocm-llvm 19.0.0.25184.60401-83~22.04 amd64 ROCm core compiler 2025-07-17T08:36:04.4398193Z ii rocm-opencl 2.0.0.60401-83~22.04 amd64 clr built using CMake 2025-07-17T08:36:04.4398554Z ii rocm-opencl-dev 2.0.0.60401-83~22.04 amd64 clr built using CMake 2025-07-17T08:36:04.4399010Z ii rocm-smi-lib 7.5.0.60401-83~22.04 amd64 AMD System Management libraries 2025-07-17T08:36:04.4399378Z ii rocm-utils 6.4.1.60401-83~22.04 amd64 Radeon Open Compute (ROCm) Runtime software stack 2025-07-17T08:36:04.4399749Z ii rocminfo 1.0.0.60401-83~22.04 amd64 Radeon Open Compute (ROCm) Runtime rocminfo tool 2025-07-17T08:36:04.4421429Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-07-17T08:36:04.4421794Z # ignore expansion of "docker ps -q" since it could be empty 2025-07-17T08:36:04.4422046Z # shellcheck disable=SC2046 2025-07-17T08:36:04.4422258Z docker stop $(docker ps -q) || true 2025-07-17T08:36:04.4422481Z # Prune all stopped containers. 2025-07-17T08:36:04.4422692Z docker container prune -f 2025-07-17T08:36:04.4435051Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:04.4435282Z env: 2025-07-17T08:36:04.4435411Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:04.4435586Z ##[endgroup] 2025-07-17T08:36:04.4794708Z docker: 'docker stop' requires at least 1 argument 2025-07-17T08:36:04.4794911Z 2025-07-17T08:36:04.4795022Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-07-17T08:36:04.4795188Z 2025-07-17T08:36:04.4795286Z See 'docker stop --help' for more information 2025-07-17T08:36:04.4943159Z Total reclaimed space: 0B 2025-07-17T08:36:04.4980675Z ##[group]Run cat /etc/os-release || true 2025-07-17T08:36:04.4980916Z cat /etc/os-release || true 2025-07-17T08:36:04.4981142Z cat /etc/apt/sources.list.d/rocm.list || true 2025-07-17T08:36:04.4981375Z cat /opt/rocm/.info/version || true 2025-07-17T08:36:04.4981585Z whoami 2025-07-17T08:36:04.5001249Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:04.5001506Z env: 2025-07-17T08:36:04.5001652Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:04.5001809Z ##[endgroup] 2025-07-17T08:36:04.5039566Z PRETTY_NAME="Ubuntu 22.04.5 LTS" 2025-07-17T08:36:04.5039755Z NAME="Ubuntu" 2025-07-17T08:36:04.5039887Z VERSION_ID="22.04" 2025-07-17T08:36:04.5040045Z VERSION="22.04.5 LTS (Jammy Jellyfish)" 2025-07-17T08:36:04.5040229Z VERSION_CODENAME=jammy 2025-07-17T08:36:04.5040384Z ID=ubuntu 2025-07-17T08:36:04.5040505Z ID_LIKE=debian 2025-07-17T08:36:04.5040682Z HOME_URL="https://www.ubuntu.com/" 2025-07-17T08:36:04.5040893Z SUPPORT_URL="https://help.ubuntu.com/" 2025-07-17T08:36:04.5041117Z BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2025-07-17T08:36:04.5041440Z PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2025-07-17T08:36:04.5041719Z UBUNTU_CODENAME=jammy 2025-07-17T08:36:04.5050268Z deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/rocm/apt/6.4.1 jammy main 2025-07-17T08:36:04.5060892Z 6.4.1-83 2025-07-17T08:36:04.5073528Z runner 2025-07-17T08:36:04.5094216Z ##[group]Run dpkg -l | grep -E " amdgpu" 2025-07-17T08:36:04.5094443Z dpkg -l | grep -E " amdgpu" 2025-07-17T08:36:04.5104492Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:04.5104715Z env: 2025-07-17T08:36:04.5104851Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:04.5105005Z ##[endgroup] 2025-07-17T08:36:04.5191251Z ii amdgpu-core 1:6.4.60401-2164967.22.04 all Core meta package for unified amdgpu driver. 2025-07-17T08:36:04.5191709Z ii amdgpu-install 6.4.60401-2164967.22.04 all AMDGPU driver repository and installer 2025-07-17T08:36:04.5216393Z ##[group]Run rocm-smi 2025-07-17T08:36:04.5216589Z rocm-smi 2025-07-17T08:36:04.5228475Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:04.5228970Z env: 2025-07-17T08:36:04.5229123Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:04.5229299Z ##[endgroup] 2025-07-17T08:36:04.7849339Z 2025-07-17T08:36:04.7849551Z 2025-07-17T08:36:04.7850053Z ============================================ ROCm System Management Interface ============================================ 2025-07-17T08:36:04.7850429Z ====================================================== Concise Info ====================================================== 2025-07-17T08:36:04.7850810Z Device Node IDs Temp Power Partitions SCLK MCLK Fan Perf PwrCap VRAM% GPU% 2025-07-17T08:36:04.7851477Z  (DID, GUID) (Junction) (Socket) (Mem, Compute, ID)  2025-07-17T08:36:04.7851799Z ========================================================================================================================== 2025-07-17T08:36:04.7852555Z 0 7 0x74a1, 26954 30.0°C 131.0W NPS1, SPX, 0 131Mhz 900Mhz 0% auto 750.0W 0% 0% 2025-07-17T08:36:04.7853018Z 1 8 0x74a1, 16738 37.0°C 135.0W NPS1, SPX, 0 134Mhz 900Mhz 0% auto 750.0W 0% 0% 2025-07-17T08:36:04.7854442Z ========================================================================================================================== 2025-07-17T08:36:04.7854741Z ================================================== End of ROCm SMI Log =================================================== 2025-07-17T08:36:04.7960174Z ##[group]Run rocminfo 2025-07-17T08:36:04.7960375Z rocminfo 2025-07-17T08:36:04.7972244Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:04.7972479Z env: 2025-07-17T08:36:04.7972627Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:04.7972786Z ##[endgroup] 2025-07-17T08:36:05.0251601Z ROCk module version 6.12.12 is loaded 2025-07-17T08:36:05.0251926Z ===================== 2025-07-17T08:36:05.0252125Z HSA System Attributes 2025-07-17T08:36:05.0252296Z ===================== 2025-07-17T08:36:05.0252971Z Runtime Version: 1.15 2025-07-17T08:36:05.0253152Z Runtime Ext Version: 1.7 2025-07-17T08:36:05.0253323Z System Timestamp Freq.: 1000.000000MHz 2025-07-17T08:36:05.0253640Z Sig. Max Wait Duration: 18446744073709551615 (0xFFFFFFFFFFFFFFFF) (timestamp count) 2025-07-17T08:36:05.0253959Z Machine Model: LARGE 2025-07-17T08:36:05.0254209Z System Endianness: LITTLE 2025-07-17T08:36:05.0254433Z Mwaitx: DISABLED 2025-07-17T08:36:05.0254604Z XNACK enabled: NO 2025-07-17T08:36:05.0254773Z DMAbuf Support: YES 2025-07-17T08:36:05.0268102Z VMM Support: YES 2025-07-17T08:36:05.0268233Z 2025-07-17T08:36:05.0268291Z ========== 2025-07-17T08:36:05.0268470Z HSA Agents 2025-07-17T08:36:05.0268623Z ========== 2025-07-17T08:36:05.0268785Z ******* 2025-07-17T08:36:05.0268959Z Agent 1 2025-07-17T08:36:05.0269128Z ******* 2025-07-17T08:36:05.0269321Z Name: AMD EPYC 9534 64-Core Processor 2025-07-17T08:36:05.0269582Z Uuid: CPU-XX 2025-07-17T08:36:05.0269822Z Marketing Name: AMD EPYC 9534 64-Core Processor 2025-07-17T08:36:05.0270059Z Vendor Name: CPU 2025-07-17T08:36:05.0270289Z Feature: None specified 2025-07-17T08:36:05.0270523Z Profile: FULL_PROFILE 2025-07-17T08:36:05.0270747Z Float Round Mode: NEAR 2025-07-17T08:36:05.0270981Z Max Queue Number: 0(0x0) 2025-07-17T08:36:05.0271200Z Queue Min Size: 0(0x0) 2025-07-17T08:36:05.0271682Z Queue Max Size: 0(0x0) 2025-07-17T08:36:05.0271904Z Queue Type: MULTI 2025-07-17T08:36:05.0272115Z Node: 0 2025-07-17T08:36:05.0272329Z Device Type: CPU 2025-07-17T08:36:05.0272532Z Cache Info: 2025-07-17T08:36:05.0272715Z L1: 32768(0x8000) KB 2025-07-17T08:36:05.0272923Z Chip ID: 0(0x0) 2025-07-17T08:36:05.0273174Z ASIC Revision: 0(0x0) 2025-07-17T08:36:05.0273401Z Cacheline Size: 64(0x40) 2025-07-17T08:36:05.0273639Z Max Clock Freq. (MHz): 2450 2025-07-17T08:36:05.0273852Z BDFID: 0 2025-07-17T08:36:05.0274191Z Internal Node ID: 0 2025-07-17T08:36:05.0274435Z Compute Unit: 128 2025-07-17T08:36:05.0274654Z SIMDs per CU: 0 2025-07-17T08:36:05.0274883Z Shader Engines: 0 2025-07-17T08:36:05.0275111Z Shader Arrs. per Eng.: 0 2025-07-17T08:36:05.0275355Z WatchPts on Addr. Ranges:1 2025-07-17T08:36:05.0275569Z Memory Properties: 2025-07-17T08:36:05.0275737Z Features: None 2025-07-17T08:36:05.0275903Z Pool Info: 2025-07-17T08:36:05.0276053Z Pool 1 2025-07-17T08:36:05.0276255Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-07-17T08:36:05.0276481Z Size: 1188703740(0x46da2dfc) KB 2025-07-17T08:36:05.0276705Z Allocatable: TRUE 2025-07-17T08:36:05.0276941Z Alloc Granule: 4KB 2025-07-17T08:36:05.0277319Z Alloc Recommended Granule:4KB 2025-07-17T08:36:05.0277570Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0277821Z Accessible by all: TRUE 2025-07-17T08:36:05.0278033Z Pool 2 2025-07-17T08:36:05.0278224Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-07-17T08:36:05.0278451Z Size: 1188703740(0x46da2dfc) KB 2025-07-17T08:36:05.0278663Z Allocatable: TRUE 2025-07-17T08:36:05.0278910Z Alloc Granule: 4KB 2025-07-17T08:36:05.0279165Z Alloc Recommended Granule:4KB 2025-07-17T08:36:05.0279419Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0279665Z Accessible by all: TRUE 2025-07-17T08:36:05.0279870Z Pool 3 2025-07-17T08:36:05.0280072Z Segment: GLOBAL; FLAGS: KERNARG, FINE GRAINED 2025-07-17T08:36:05.0280286Z Size: 1188703740(0x46da2dfc) KB 2025-07-17T08:36:05.0280508Z Allocatable: TRUE 2025-07-17T08:36:05.0280736Z Alloc Granule: 4KB 2025-07-17T08:36:05.0280980Z Alloc Recommended Granule:4KB 2025-07-17T08:36:05.0281226Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0281455Z Accessible by all: TRUE 2025-07-17T08:36:05.0281664Z Pool 4 2025-07-17T08:36:05.0281846Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-07-17T08:36:05.0282068Z Size: 1188703740(0x46da2dfc) KB 2025-07-17T08:36:05.0282425Z Allocatable: TRUE 2025-07-17T08:36:05.0282661Z Alloc Granule: 4KB 2025-07-17T08:36:05.0282903Z Alloc Recommended Granule:4KB 2025-07-17T08:36:05.0283142Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0283378Z Accessible by all: TRUE 2025-07-17T08:36:05.0283580Z ISA Info: 2025-07-17T08:36:05.0283738Z ******* 2025-07-17T08:36:05.0283886Z Agent 2 2025-07-17T08:36:05.0284034Z ******* 2025-07-17T08:36:05.0284202Z Name: AMD EPYC 9534 64-Core Processor 2025-07-17T08:36:05.0284503Z Uuid: CPU-XX 2025-07-17T08:36:05.0284733Z Marketing Name: AMD EPYC 9534 64-Core Processor 2025-07-17T08:36:05.0284972Z Vendor Name: CPU 2025-07-17T08:36:05.0285199Z Feature: None specified 2025-07-17T08:36:05.0285418Z Profile: FULL_PROFILE 2025-07-17T08:36:05.0285654Z Float Round Mode: NEAR 2025-07-17T08:36:05.0285882Z Max Queue Number: 0(0x0) 2025-07-17T08:36:05.0286123Z Queue Min Size: 0(0x0) 2025-07-17T08:36:05.0286350Z Queue Max Size: 0(0x0) 2025-07-17T08:36:05.0286565Z Queue Type: MULTI 2025-07-17T08:36:05.0286774Z Node: 1 2025-07-17T08:36:05.0286979Z Device Type: CPU 2025-07-17T08:36:05.0287180Z Cache Info: 2025-07-17T08:36:05.0287344Z L1: 32768(0x8000) KB 2025-07-17T08:36:05.0287557Z Chip ID: 0(0x0) 2025-07-17T08:36:05.0287887Z ASIC Revision: 0(0x0) 2025-07-17T08:36:05.0288111Z Cacheline Size: 64(0x40) 2025-07-17T08:36:05.0288342Z Max Clock Freq. (MHz): 2450 2025-07-17T08:36:05.0288550Z BDFID: 0 2025-07-17T08:36:05.0288764Z Internal Node ID: 1 2025-07-17T08:36:05.0288990Z Compute Unit: 128 2025-07-17T08:36:05.0289218Z SIMDs per CU: 0 2025-07-17T08:36:05.0289451Z Shader Engines: 0 2025-07-17T08:36:05.0289688Z Shader Arrs. per Eng.: 0 2025-07-17T08:36:05.0289943Z WatchPts on Addr. Ranges:1 2025-07-17T08:36:05.0290161Z Memory Properties: 2025-07-17T08:36:05.0290327Z Features: None 2025-07-17T08:36:05.0290501Z Pool Info: 2025-07-17T08:36:05.0290668Z Pool 1 2025-07-17T08:36:05.0290863Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-07-17T08:36:05.0291102Z Size: 1188946684(0x46dde2fc) KB 2025-07-17T08:36:05.0291322Z Allocatable: TRUE 2025-07-17T08:36:05.0291546Z Alloc Granule: 4KB 2025-07-17T08:36:05.0291787Z Alloc Recommended Granule:4KB 2025-07-17T08:36:05.0292031Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0292269Z Accessible by all: TRUE 2025-07-17T08:36:05.0292471Z Pool 2 2025-07-17T08:36:05.0292778Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-07-17T08:36:05.0293006Z Size: 1188946684(0x46dde2fc) KB 2025-07-17T08:36:05.0293222Z Allocatable: TRUE 2025-07-17T08:36:05.0293454Z Alloc Granule: 4KB 2025-07-17T08:36:05.0293704Z Alloc Recommended Granule:4KB 2025-07-17T08:36:05.0293944Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0294171Z Accessible by all: TRUE 2025-07-17T08:36:05.0294373Z Pool 3 2025-07-17T08:36:05.0294571Z Segment: GLOBAL; FLAGS: KERNARG, FINE GRAINED 2025-07-17T08:36:05.0294784Z Size: 1188946684(0x46dde2fc) KB 2025-07-17T08:36:05.0295002Z Allocatable: TRUE 2025-07-17T08:36:05.0295229Z Alloc Granule: 4KB 2025-07-17T08:36:05.0295473Z Alloc Recommended Granule:4KB 2025-07-17T08:36:05.0295710Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0295958Z Accessible by all: TRUE 2025-07-17T08:36:05.0296160Z Pool 4 2025-07-17T08:36:05.0296345Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-07-17T08:36:05.0296564Z Size: 1188946684(0x46dde2fc) KB 2025-07-17T08:36:05.0296776Z Allocatable: TRUE 2025-07-17T08:36:05.0297007Z Alloc Granule: 4KB 2025-07-17T08:36:05.0297239Z Alloc Recommended Granule:4KB 2025-07-17T08:36:05.0297480Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0297713Z Accessible by all: TRUE 2025-07-17T08:36:05.0297924Z ISA Info: 2025-07-17T08:36:05.0298227Z ******* 2025-07-17T08:36:05.0298375Z Agent 3 2025-07-17T08:36:05.0298522Z ******* 2025-07-17T08:36:05.0298681Z Name: gfx942 2025-07-17T08:36:05.0298899Z Uuid: GPU-6d96a91a4aa636e8 2025-07-17T08:36:05.0299122Z Marketing Name: AMD Instinct MI300X 2025-07-17T08:36:05.0299353Z Vendor Name: AMD 2025-07-17T08:36:05.0299575Z Feature: KERNEL_DISPATCH 2025-07-17T08:36:05.0299800Z Profile: BASE_PROFILE 2025-07-17T08:36:05.0300028Z Float Round Mode: NEAR 2025-07-17T08:36:05.0300256Z Max Queue Number: 128(0x80) 2025-07-17T08:36:05.0300484Z Queue Min Size: 64(0x40) 2025-07-17T08:36:05.0300706Z Queue Max Size: 131072(0x20000) 2025-07-17T08:36:05.0300930Z Queue Type: MULTI 2025-07-17T08:36:05.0301135Z Node: 2 2025-07-17T08:36:05.0301350Z Device Type: GPU 2025-07-17T08:36:05.0301550Z Cache Info: 2025-07-17T08:36:05.0301719Z L1: 32(0x20) KB 2025-07-17T08:36:05.0301921Z L2: 4096(0x1000) KB 2025-07-17T08:36:05.0302104Z L3: 262144(0x40000) KB 2025-07-17T08:36:05.0302306Z Chip ID: 29857(0x74a1) 2025-07-17T08:36:05.0302522Z ASIC Revision: 1(0x1) 2025-07-17T08:36:05.0302753Z Cacheline Size: 128(0x80) 2025-07-17T08:36:05.0303125Z Max Clock Freq. (MHz): 2100 2025-07-17T08:36:05.0303341Z BDFID: 42752 2025-07-17T08:36:05.0303560Z Internal Node ID: 2 2025-07-17T08:36:05.0303787Z Compute Unit: 304 2025-07-17T08:36:05.0304006Z SIMDs per CU: 4 2025-07-17T08:36:05.0304229Z Shader Engines: 32 2025-07-17T08:36:05.0304470Z Shader Arrs. per Eng.: 1 2025-07-17T08:36:05.0304713Z WatchPts on Addr. Ranges:4 2025-07-17T08:36:05.0304951Z Coherent Host Access: FALSE 2025-07-17T08:36:05.0305169Z Memory Properties: 2025-07-17T08:36:05.0305341Z Features: KERNEL_DISPATCH 2025-07-17T08:36:05.0305566Z Fast F16 Operation: TRUE 2025-07-17T08:36:05.0305795Z Wavefront Size: 64(0x40) 2025-07-17T08:36:05.0306030Z Workgroup Max Size: 1024(0x400) 2025-07-17T08:36:05.0306247Z Workgroup Max Size per Dimension: 2025-07-17T08:36:05.0306429Z x 1024(0x400) 2025-07-17T08:36:05.0306620Z y 1024(0x400) 2025-07-17T08:36:05.0306800Z z 1024(0x400) 2025-07-17T08:36:05.0307002Z Max Waves Per CU: 32(0x20) 2025-07-17T08:36:05.0307228Z Max Work-item Per CU: 2048(0x800) 2025-07-17T08:36:05.0307457Z Grid Max Size: 4294967295(0xffffffff) 2025-07-17T08:36:05.0307664Z Grid Max Size per Dimension: 2025-07-17T08:36:05.0307827Z x 4294967295(0xffffffff) 2025-07-17T08:36:05.0308024Z y 4294967295(0xffffffff) 2025-07-17T08:36:05.0308320Z z 4294967295(0xffffffff) 2025-07-17T08:36:05.0308547Z Max fbarriers/Workgrp: 32 2025-07-17T08:36:05.0314791Z Packet Processor uCode:: 177 2025-07-17T08:36:05.0315085Z SDMA engine uCode:: 24 2025-07-17T08:36:05.0315339Z IOMMU Support:: None 2025-07-17T08:36:05.0315546Z Pool Info: 2025-07-17T08:36:05.0315711Z Pool 1 2025-07-17T08:36:05.0315908Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-07-17T08:36:05.0316148Z Size: 201310208(0xbffc000) KB 2025-07-17T08:36:05.0316386Z Allocatable: TRUE 2025-07-17T08:36:05.0316621Z Alloc Granule: 4KB 2025-07-17T08:36:05.0316877Z Alloc Recommended Granule:2048KB 2025-07-17T08:36:05.0317138Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0317393Z Accessible by all: FALSE 2025-07-17T08:36:05.0317603Z Pool 2 2025-07-17T08:36:05.0317806Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-07-17T08:36:05.0318043Z Size: 201310208(0xbffc000) KB 2025-07-17T08:36:05.0318268Z Allocatable: TRUE 2025-07-17T08:36:05.0318505Z Alloc Granule: 4KB 2025-07-17T08:36:05.0318742Z Alloc Recommended Granule:2048KB 2025-07-17T08:36:05.0318990Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0319222Z Accessible by all: FALSE 2025-07-17T08:36:05.0319620Z Pool 3 2025-07-17T08:36:05.0319810Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-07-17T08:36:05.0320031Z Size: 201310208(0xbffc000) KB 2025-07-17T08:36:05.0320247Z Allocatable: TRUE 2025-07-17T08:36:05.0320480Z Alloc Granule: 4KB 2025-07-17T08:36:05.0320722Z Alloc Recommended Granule:2048KB 2025-07-17T08:36:05.0320964Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0321204Z Accessible by all: FALSE 2025-07-17T08:36:05.0321406Z Pool 4 2025-07-17T08:36:05.0321590Z Segment: GROUP 2025-07-17T08:36:05.0321797Z Size: 64(0x40) KB 2025-07-17T08:36:05.0322021Z Allocatable: FALSE 2025-07-17T08:36:05.0322255Z Alloc Granule: 0KB 2025-07-17T08:36:05.0322494Z Alloc Recommended Granule:0KB 2025-07-17T08:36:05.0322735Z Alloc Alignment: 0KB 2025-07-17T08:36:05.0322964Z Accessible by all: FALSE 2025-07-17T08:36:05.0323169Z ISA Info: 2025-07-17T08:36:05.0323321Z ISA 1 2025-07-17T08:36:05.0323518Z Name: amdgcn-amd-amdhsa--gfx942:sramecc+:xnack- 2025-07-17T08:36:05.0323771Z Machine Models: HSA_MACHINE_MODEL_LARGE 2025-07-17T08:36:05.0324007Z Profiles: HSA_PROFILE_BASE 2025-07-17T08:36:05.0324252Z Default Rounding Mode: NEAR 2025-07-17T08:36:05.0324492Z Default Rounding Mode: NEAR 2025-07-17T08:36:05.0324727Z Fast f16: TRUE 2025-07-17T08:36:05.0325082Z Workgroup Max Size: 1024(0x400) 2025-07-17T08:36:05.0325314Z Workgroup Max Size per Dimension: 2025-07-17T08:36:05.0325518Z x 1024(0x400) 2025-07-17T08:36:05.0325709Z y 1024(0x400) 2025-07-17T08:36:05.0325894Z z 1024(0x400) 2025-07-17T08:36:05.0326100Z Grid Max Size: 4294967295(0xffffffff) 2025-07-17T08:36:05.0326317Z Grid Max Size per Dimension: 2025-07-17T08:36:05.0326489Z x 4294967295(0xffffffff) 2025-07-17T08:36:05.0326682Z y 4294967295(0xffffffff) 2025-07-17T08:36:05.0326874Z z 4294967295(0xffffffff) 2025-07-17T08:36:05.0327086Z FBarrier Max Size: 32 2025-07-17T08:36:05.0327294Z ISA 2 2025-07-17T08:36:05.0327507Z Name: amdgcn-amd-amdhsa--gfx9-4-generic:sramecc+:xnack- 2025-07-17T08:36:05.0327772Z Machine Models: HSA_MACHINE_MODEL_LARGE 2025-07-17T08:36:05.0328006Z Profiles: HSA_PROFILE_BASE 2025-07-17T08:36:05.0328246Z Default Rounding Mode: NEAR 2025-07-17T08:36:05.0328491Z Default Rounding Mode: NEAR 2025-07-17T08:36:05.0328711Z Fast f16: TRUE 2025-07-17T08:36:05.0328941Z Workgroup Max Size: 1024(0x400) 2025-07-17T08:36:05.0329157Z Workgroup Max Size per Dimension: 2025-07-17T08:36:05.0329342Z x 1024(0x400) 2025-07-17T08:36:05.0329638Z y 1024(0x400) 2025-07-17T08:36:05.0329831Z z 1024(0x400) 2025-07-17T08:36:05.0330039Z Grid Max Size: 4294967295(0xffffffff) 2025-07-17T08:36:05.0330245Z Grid Max Size per Dimension: 2025-07-17T08:36:05.0330421Z x 4294967295(0xffffffff) 2025-07-17T08:36:05.0330605Z y 4294967295(0xffffffff) 2025-07-17T08:36:05.0330797Z z 4294967295(0xffffffff) 2025-07-17T08:36:05.0331000Z FBarrier Max Size: 32 2025-07-17T08:36:05.0331198Z ******* 2025-07-17T08:36:05.0331341Z Agent 4 2025-07-17T08:36:05.0331484Z ******* 2025-07-17T08:36:05.0331648Z Name: gfx942 2025-07-17T08:36:05.0331861Z Uuid: GPU-e07d565caf047d31 2025-07-17T08:36:05.0332092Z Marketing Name: AMD Instinct MI300X 2025-07-17T08:36:05.0332323Z Vendor Name: AMD 2025-07-17T08:36:05.0332549Z Feature: KERNEL_DISPATCH 2025-07-17T08:36:05.0332785Z Profile: BASE_PROFILE 2025-07-17T08:36:05.0333021Z Float Round Mode: NEAR 2025-07-17T08:36:05.0333248Z Max Queue Number: 128(0x80) 2025-07-17T08:36:05.0333473Z Queue Min Size: 64(0x40) 2025-07-17T08:36:05.0333695Z Queue Max Size: 131072(0x20000) 2025-07-17T08:36:05.0333909Z Queue Type: MULTI 2025-07-17T08:36:05.0334115Z Node: 3 2025-07-17T08:36:05.0334324Z Device Type: GPU 2025-07-17T08:36:05.0334527Z Cache Info: 2025-07-17T08:36:05.0334806Z L1: 32(0x20) KB 2025-07-17T08:36:05.0335010Z L2: 4096(0x1000) KB 2025-07-17T08:36:05.0335200Z L3: 262144(0x40000) KB 2025-07-17T08:36:05.0335393Z Chip ID: 29857(0x74a1) 2025-07-17T08:36:05.0335618Z ASIC Revision: 1(0x1) 2025-07-17T08:36:05.0335845Z Cacheline Size: 128(0x80) 2025-07-17T08:36:05.0336075Z Max Clock Freq. (MHz): 2100 2025-07-17T08:36:05.0336285Z BDFID: 50944 2025-07-17T08:36:05.0336504Z Internal Node ID: 3 2025-07-17T08:36:05.0336730Z Compute Unit: 304 2025-07-17T08:36:05.0336951Z SIMDs per CU: 4 2025-07-17T08:36:05.0337180Z Shader Engines: 32 2025-07-17T08:36:05.0337407Z Shader Arrs. per Eng.: 1 2025-07-17T08:36:05.0337648Z WatchPts on Addr. Ranges:4 2025-07-17T08:36:05.0337887Z Coherent Host Access: FALSE 2025-07-17T08:36:05.0338103Z Memory Properties: 2025-07-17T08:36:05.0338277Z Features: KERNEL_DISPATCH 2025-07-17T08:36:05.0338492Z Fast F16 Operation: TRUE 2025-07-17T08:36:05.0338728Z Wavefront Size: 64(0x40) 2025-07-17T08:36:05.0338958Z Workgroup Max Size: 1024(0x400) 2025-07-17T08:36:05.0339175Z Workgroup Max Size per Dimension: 2025-07-17T08:36:05.0339358Z x 1024(0x400) 2025-07-17T08:36:05.0339668Z y 1024(0x400) 2025-07-17T08:36:05.0339854Z z 1024(0x400) 2025-07-17T08:36:05.0340062Z Max Waves Per CU: 32(0x20) 2025-07-17T08:36:05.0340293Z Max Work-item Per CU: 2048(0x800) 2025-07-17T08:36:05.0340519Z Grid Max Size: 4294967295(0xffffffff) 2025-07-17T08:36:05.0340729Z Grid Max Size per Dimension: 2025-07-17T08:36:05.0340891Z x 4294967295(0xffffffff) 2025-07-17T08:36:05.0341086Z y 4294967295(0xffffffff) 2025-07-17T08:36:05.0341266Z z 4294967295(0xffffffff) 2025-07-17T08:36:05.0341483Z Max fbarriers/Workgrp: 32 2025-07-17T08:36:05.0341733Z Packet Processor uCode:: 177 2025-07-17T08:36:05.0341976Z SDMA engine uCode:: 24 2025-07-17T08:36:05.0342218Z IOMMU Support:: None 2025-07-17T08:36:05.0342415Z Pool Info: 2025-07-17T08:36:05.0342571Z Pool 1 2025-07-17T08:36:05.0342764Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-07-17T08:36:05.0342997Z Size: 201310208(0xbffc000) KB 2025-07-17T08:36:05.0343224Z Allocatable: TRUE 2025-07-17T08:36:05.0343455Z Alloc Granule: 4KB 2025-07-17T08:36:05.0343707Z Alloc Recommended Granule:2048KB 2025-07-17T08:36:05.0343948Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0344190Z Accessible by all: FALSE 2025-07-17T08:36:05.0344392Z Pool 2 2025-07-17T08:36:05.0344583Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-07-17T08:36:05.0344943Z Size: 201310208(0xbffc000) KB 2025-07-17T08:36:05.0345162Z Allocatable: TRUE 2025-07-17T08:36:05.0345397Z Alloc Granule: 4KB 2025-07-17T08:36:05.0345634Z Alloc Recommended Granule:2048KB 2025-07-17T08:36:05.0345877Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0346105Z Accessible by all: FALSE 2025-07-17T08:36:05.0346309Z Pool 3 2025-07-17T08:36:05.0346493Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-07-17T08:36:05.0346704Z Size: 201310208(0xbffc000) KB 2025-07-17T08:36:05.0346923Z Allocatable: TRUE 2025-07-17T08:36:05.0347154Z Alloc Granule: 4KB 2025-07-17T08:36:05.0347396Z Alloc Recommended Granule:2048KB 2025-07-17T08:36:05.0347634Z Alloc Alignment: 4KB 2025-07-17T08:36:05.0347868Z Accessible by all: FALSE 2025-07-17T08:36:05.0348067Z Pool 4 2025-07-17T08:36:05.0348241Z Segment: GROUP 2025-07-17T08:36:05.0348445Z Size: 64(0x40) KB 2025-07-17T08:36:05.0348654Z Allocatable: FALSE 2025-07-17T08:36:05.0348887Z Alloc Granule: 0KB 2025-07-17T08:36:05.0349121Z Alloc Recommended Granule:0KB 2025-07-17T08:36:05.0349362Z Alloc Alignment: 0KB 2025-07-17T08:36:05.0357633Z Accessible by all: FALSE 2025-07-17T08:36:05.0357846Z ISA Info: 2025-07-17T08:36:05.0357999Z ISA 1 2025-07-17T08:36:05.0358189Z Name: amdgcn-amd-amdhsa--gfx942:sramecc+:xnack- 2025-07-17T08:36:05.0358445Z Machine Models: HSA_MACHINE_MODEL_LARGE 2025-07-17T08:36:05.0358683Z Profiles: HSA_PROFILE_BASE 2025-07-17T08:36:05.0358924Z Default Rounding Mode: NEAR 2025-07-17T08:36:05.0359160Z Default Rounding Mode: NEAR 2025-07-17T08:36:05.0359384Z Fast f16: TRUE 2025-07-17T08:36:05.0359608Z Workgroup Max Size: 1024(0x400) 2025-07-17T08:36:05.0359830Z Workgroup Max Size per Dimension: 2025-07-17T08:36:05.0360017Z x 1024(0x400) 2025-07-17T08:36:05.0360212Z y 1024(0x400) 2025-07-17T08:36:05.0360401Z z 1024(0x400) 2025-07-17T08:36:05.0360612Z Grid Max Size: 4294967295(0xffffffff) 2025-07-17T08:36:05.0360828Z Grid Max Size per Dimension: 2025-07-17T08:36:05.0361007Z x 4294967295(0xffffffff) 2025-07-17T08:36:05.0361205Z y 4294967295(0xffffffff) 2025-07-17T08:36:05.0361393Z z 4294967295(0xffffffff) 2025-07-17T08:36:05.0361600Z FBarrier Max Size: 32 2025-07-17T08:36:05.0361830Z ISA 2 2025-07-17T08:36:05.0362035Z Name: amdgcn-amd-amdhsa--gfx9-4-generic:sramecc+:xnack- 2025-07-17T08:36:05.0362289Z Machine Models: HSA_MACHINE_MODEL_LARGE 2025-07-17T08:36:05.0362532Z Profiles: HSA_PROFILE_BASE 2025-07-17T08:36:05.0362925Z Default Rounding Mode: NEAR 2025-07-17T08:36:05.0363181Z Default Rounding Mode: NEAR 2025-07-17T08:36:05.0363400Z Fast f16: TRUE 2025-07-17T08:36:05.0363618Z Workgroup Max Size: 1024(0x400) 2025-07-17T08:36:05.0363844Z Workgroup Max Size per Dimension: 2025-07-17T08:36:05.0364026Z x 1024(0x400) 2025-07-17T08:36:05.0364219Z y 1024(0x400) 2025-07-17T08:36:05.0364402Z z 1024(0x400) 2025-07-17T08:36:05.0364611Z Grid Max Size: 4294967295(0xffffffff) 2025-07-17T08:36:05.0364815Z Grid Max Size per Dimension: 2025-07-17T08:36:05.0364998Z x 4294967295(0xffffffff) 2025-07-17T08:36:05.0365195Z y 4294967295(0xffffffff) 2025-07-17T08:36:05.0365379Z z 4294967295(0xffffffff) 2025-07-17T08:36:05.0365594Z FBarrier Max Size: 32 2025-07-17T08:36:05.0365795Z *** Done *** 2025-07-17T08:36:05.0426985Z ##[group]Run ngpu=$(rocminfo | grep -c -E 'Name:.*\sgfx') 2025-07-17T08:36:05.0427297Z ngpu=$(rocminfo | grep -c -E 'Name:.*\sgfx') 2025-07-17T08:36:05.0427756Z msg="Please file an issue on pytorch/pytorch reporting the faulty runner. Include a link to the runner logs so the runner can be identified" 2025-07-17T08:36:05.0428192Z if [[ $ngpu -eq 0 ]]; then 2025-07-17T08:36:05.0428423Z  echo "Error: Failed to detect any GPUs on the runner" 2025-07-17T08:36:05.0428645Z  echo "$msg" 2025-07-17T08:36:05.0429061Z  exit 1 2025-07-17T08:36:05.0429207Z fi 2025-07-17T08:36:05.0429358Z if [[ $ngpu -eq 1 ]]; then 2025-07-17T08:36:05.0429623Z  echo "Error: only 1 GPU detected, at least 2 GPUs are needed for distributed jobs" 2025-07-17T08:36:05.0429899Z  echo "$msg" 2025-07-17T08:36:05.0430047Z  exit 1 2025-07-17T08:36:05.0430177Z fi 2025-07-17T08:36:05.0442061Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:05.0442297Z env: 2025-07-17T08:36:05.0442442Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:05.0442601Z ##[endgroup] 2025-07-17T08:36:05.3894001Z ##[group]Run pytorch/pytorch/.github/actions/diskspace-cleanup@main 2025-07-17T08:36:05.3894379Z with: 2025-07-17T08:36:05.3894569Z diskspace-cutoff: 70 2025-07-17T08:36:05.3894748Z env: 2025-07-17T08:36:05.3894874Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:05.3895039Z ##[endgroup] 2025-07-17T08:36:05.3927381Z ##[group]Run set -ex 2025-07-17T08:36:05.3927660Z set -ex 2025-07-17T08:36:05.3927886Z diskspace_cutoff=70 2025-07-17T08:36:05.3928203Z docker_root_dir=$(docker info -f '{{.DockerRootDir}}') 2025-07-17T08:36:05.3928458Z if [ ! -d "$docker_root_dir" ]; then 2025-07-17T08:36:05.3928782Z  echo "Docker root directory ($docker_root_dir) does not exist. Skipping disk space check." 2025-07-17T08:36:05.3929079Z  exit 0 2025-07-17T08:36:05.3929234Z fi 2025-07-17T08:36:05.3929498Z diskspace=$(df -H --output=pcent ${docker_root_dir} | sed -n 2p | sed 's/%//' | sed 's/ //') 2025-07-17T08:36:05.3930052Z msg="Please file an issue on pytorch/pytorch reporting the faulty runner. Include a link to the runner logs so the runner can be identified" 2025-07-17T08:36:05.3930509Z if [[ "$diskspace" -ge "$diskspace_cutoff" ]] ; then 2025-07-17T08:36:05.3930740Z  docker system prune -af 2025-07-17T08:36:05.3931050Z  diskspace_new=$(df -H --output=pcent ${docker_root_dir} | sed -n 2p | sed 's/%//' | sed 's/ //') 2025-07-17T08:36:05.3931599Z  if [[ "$diskspace_new" -gt "$diskspace_cutoff" ]] ; then 2025-07-17T08:36:05.3931964Z  echo "Error: Available diskspace is less than $diskspace_cutoff percent. Not enough diskspace." 2025-07-17T08:36:05.3932283Z  echo "$msg" 2025-07-17T08:36:05.3932447Z  exit 1 2025-07-17T08:36:05.3932602Z  else 2025-07-17T08:36:05.3932778Z  difference=$((diskspace - diskspace_new)) 2025-07-17T08:36:05.3933029Z  echo "Diskspace saved: $difference percent" 2025-07-17T08:36:05.3933230Z  fi 2025-07-17T08:36:05.3933367Z fi 2025-07-17T08:36:05.3945757Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:05.3946094Z env: 2025-07-17T08:36:05.3946271Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:05.3946442Z ##[endgroup] 2025-07-17T08:36:05.3979795Z + diskspace_cutoff=70 2025-07-17T08:36:05.3984048Z ++ docker info -f '{{.DockerRootDir}}' 2025-07-17T08:36:05.4587337Z + docker_root_dir=/home/runner/docker-data 2025-07-17T08:36:05.4587628Z + '[' '!' -d /home/runner/docker-data ']' 2025-07-17T08:36:05.4595791Z ++ df -H --output=pcent /home/runner/docker-data 2025-07-17T08:36:05.4596413Z ++ sed -n 2p 2025-07-17T08:36:05.4597084Z ++ sed s/%// 2025-07-17T08:36:05.4599162Z ++ sed 's/ //' 2025-07-17T08:36:05.4615935Z + diskspace=' 1' 2025-07-17T08:36:05.4616561Z + msg='Please file an issue on pytorch/pytorch reporting the faulty runner. Include a link to the runner logs so the runner can be identified' 2025-07-17T08:36:05.4616983Z + [[ 1 -ge 70 ]] 2025-07-17T08:36:05.4647916Z ##[group]Run RUNNER_ARTIFACT_DIR="${RUNNER_TEMP}/artifacts" 2025-07-17T08:36:05.4648229Z RUNNER_ARTIFACT_DIR="${RUNNER_TEMP}/artifacts" 2025-07-17T08:36:05.4648463Z rm -rf "${RUNNER_ARTIFACT_DIR}" 2025-07-17T08:36:05.4648689Z mkdir -p "${RUNNER_ARTIFACT_DIR}" 2025-07-17T08:36:05.4649286Z echo "RUNNER_ARTIFACT_DIR=${RUNNER_ARTIFACT_DIR}" >> "${GITHUB_ENV}" 2025-07-17T08:36:05.4649555Z  2025-07-17T08:36:05.4649761Z RUNNER_TEST_RESULTS_DIR="${RUNNER_TEMP}/test-results" 2025-07-17T08:36:05.4650021Z rm -rf "${RUNNER_TEST_RESULTS_DIR}" 2025-07-17T08:36:05.4650246Z mkdir -p "${RUNNER_TEST_RESULTS_DIR}" 2025-07-17T08:36:05.4650531Z echo "RUNNER_TEST_RESULTS_DIR=${RUNNER_TEST_RESULTS_DIR}" >> "${GITHUB_ENV}" 2025-07-17T08:36:05.4650813Z  2025-07-17T08:36:05.4650966Z RUNNER_DOCS_DIR="${RUNNER_TEMP}/docs" 2025-07-17T08:36:05.4651173Z rm -rf "${RUNNER_DOCS_DIR}" 2025-07-17T08:36:05.4651372Z mkdir -p "${RUNNER_DOCS_DIR}" 2025-07-17T08:36:05.4651619Z echo "RUNNER_DOCS_DIR=${RUNNER_DOCS_DIR}" >> "${GITHUB_ENV}" 2025-07-17T08:36:05.4663071Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:05.4663324Z env: 2025-07-17T08:36:05.4663475Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:05.4663647Z ##[endgroup] 2025-07-17T08:36:05.4779369Z ##[group]Run env | grep '^GITHUB' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-07-17T08:36:05.4779735Z env | grep '^GITHUB' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-07-17T08:36:05.4780049Z env | grep '^CI' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-07-17T08:36:05.4790319Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:05.4790580Z env: 2025-07-17T08:36:05.4790742Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:05.4790973Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:36:05.4791263Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:36:05.4791544Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:36:05.4791765Z ##[endgroup] 2025-07-17T08:36:05.4882865Z ##[group]Run # All GPUs are visible to the runner; visibility, if needed, will be set by run_test.py. 2025-07-17T08:36:05.4883518Z # All GPUs are visible to the runner; visibility, if needed, will be set by run_test.py. 2025-07-17T08:36:05.4884085Z # Add render group for container creation. 2025-07-17T08:36:05.4884359Z render_gid=`cat /etc/group | grep render | cut -d: -f3` 2025-07-17T08:36:05.4884687Z # Ensure GPU isolation if pod is part of kubernetes setup with DEVICE_FLAG. 2025-07-17T08:36:05.4885006Z if [ -f "/etc/podinfo/gha-render-devices" ]; then 2025-07-17T08:36:05.4885277Z  DEVICE_FLAG=$(cat /etc/podinfo/gha-render-devices) 2025-07-17T08:36:05.4885492Z else 2025-07-17T08:36:05.4885648Z  DEVICE_FLAG="--device /dev/dri" 2025-07-17T08:36:05.4885825Z fi 2025-07-17T08:36:05.4886116Z # The --group-add daemon and --group-add bin are needed in the Ubuntu 24.04 and Almalinux OSs respectively. 2025-07-17T08:36:05.4886554Z # This is due to the device files (/dev/kfd & /dev/dri) being owned by video group on bare metal. 2025-07-17T08:36:05.4886974Z # This video group ID maps to subgid 1 inside the docker image due to the /etc/subgid entries. 2025-07-17T08:36:05.4887401Z # The group name corresponding to group ID 1 can change depending on the OS, so both are necessary. 2025-07-17T08:36:05.4888121Z echo "GPU_FLAG=--device=/dev/mem --device=/dev/kfd $DEVICE_FLAG --group-add video --group-add $render_gid --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host" >> "${GITHUB_ENV}" 2025-07-17T08:36:05.4898815Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:05.4899142Z env: 2025-07-17T08:36:05.4899338Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:05.4899544Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:36:05.4899814Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:36:05.4900073Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:36:05.4900431Z ##[endgroup] 2025-07-17T08:36:05.5017798Z ##[group]Run aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722 2025-07-17T08:36:05.5018133Z with: 2025-07-17T08:36:05.5018362Z role-to-assume: arn:aws:iam::308535385114:role/gha_workflow_s3_and_ecr_read_only 2025-07-17T08:36:05.5018641Z aws-region: us-east-1 2025-07-17T08:36:05.5018811Z role-duration-seconds: 18000 2025-07-17T08:36:05.5018993Z audience: sts.amazonaws.com 2025-07-17T08:36:05.5019162Z env: 2025-07-17T08:36:05.5019290Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:05.5019510Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:36:05.5019785Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:36:05.5020043Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:36:05.5020787Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:36:05.5021453Z ##[endgroup] 2025-07-17T08:36:05.7194430Z Assuming role with OIDC 2025-07-17T08:36:07.8550513Z Authenticated as assumedRoleId AROAUPVRELQNLLCOPFEJR:GitHubActions 2025-07-17T08:36:07.9015421Z ##[group]Run aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076 2025-07-17T08:36:07.9015742Z with: 2025-07-17T08:36:07.9015898Z mask-password: true 2025-07-17T08:36:07.9016072Z registry-type: private 2025-07-17T08:36:07.9016250Z skip-logout: false 2025-07-17T08:36:07.9016400Z env: 2025-07-17T08:36:07.9016550Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:07.9016773Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:36:07.9017067Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:36:07.9017347Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:36:07.9018057Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:36:07.9018746Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:36:07.9018971Z AWS_REGION: us-east-1 2025-07-17T08:36:07.9019478Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:36:07.9019758Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:36:07.9023571Z AWS_SESSION_TOKEN: *** 2025-07-17T08:36:07.9023755Z ##[endgroup] 2025-07-17T08:36:10.1511519Z Logging into registry 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:12.3963707Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-07-17T08:36:12.3964022Z with: 2025-07-17T08:36:12.3964477Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:12.3964970Z use-custom-docker-registry: true 2025-07-17T08:36:12.3965197Z docker-build-dir: .ci/docker 2025-07-17T08:36:12.3965388Z docker-build-script: ./build.sh 2025-07-17T08:36:12.3965581Z working-directory: . 2025-07-17T08:36:12.3965809Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:12.3966061Z force-push: false 2025-07-17T08:36:12.3966220Z env: 2025-07-17T08:36:12.3966364Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:12.3966584Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:36:12.3966870Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:36:12.3967177Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:36:12.3967863Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:36:12.3968520Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:36:12.3968915Z AWS_REGION: us-east-1 2025-07-17T08:36:12.3969228Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:36:12.3969482Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:36:12.3973276Z AWS_SESSION_TOKEN: *** 2025-07-17T08:36:12.3973445Z ##[endgroup] 2025-07-17T08:36:12.3991332Z ##[group]Run set -ex 2025-07-17T08:36:12.3991534Z set -ex 2025-07-17T08:36:12.3991684Z  2025-07-17T08:36:12.3991935Z # If the docker build directory or the build script doesn't exist, the action will 2025-07-17T08:36:12.3992344Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-07-17T08:36:12.3992684Z # job could then download the pre-built image as usual 2025-07-17T08:36:12.3993103Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-07-17T08:36:12.3993487Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-07-17T08:36:12.3993692Z else 2025-07-17T08:36:12.3993870Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-07-17T08:36:12.3994239Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-07-17T08:36:12.3994488Z  2025-07-17T08:36:12.3994822Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-07-17T08:36:12.3995187Z  exit 0 2025-07-17T08:36:12.3995333Z fi 2025-07-17T08:36:12.3995472Z  2025-07-17T08:36:12.3995692Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-07-17T08:36:12.3996051Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-07-17T08:36:12.3996374Z  # use it as it is, but first let's extract the tag 2025-07-17T08:36:12.3996670Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-07-17T08:36:12.3996991Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-07-17T08:36:12.3997294Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-07-17T08:36:12.3997542Z else 2025-07-17T08:36:12.3997712Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-07-17T08:36:12.3997954Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-07-17T08:36:12.3998205Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-07-17T08:36:12.3998410Z  fi 2025-07-17T08:36:12.4004987Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-07-17T08:36:12.4005389Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-07-17T08:36:12.4005771Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-07-17T08:36:12.4006181Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-07-17T08:36:12.4006441Z fi 2025-07-17T08:36:12.4018589Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:12.4018828Z env: 2025-07-17T08:36:12.4018981Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:12.4019202Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:36:12.4019492Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:36:12.4019760Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:36:12.4020455Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:36:12.4021119Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:36:12.4021307Z AWS_REGION: us-east-1 2025-07-17T08:36:12.4021538Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:36:12.4021792Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:36:12.4025601Z AWS_SESSION_TOKEN: *** 2025-07-17T08:36:12.4025943Z REPO_NAME: pytorch 2025-07-17T08:36:12.4026413Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:12.4026887Z DOCKER_BUILD_DIR: .ci/docker 2025-07-17T08:36:12.4027086Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-07-17T08:36:12.4027331Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:12.4027586Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-07-17T08:36:12.4027777Z CUSTOM_TAG_PREFIX: 2025-07-17T08:36:12.4027935Z ##[endgroup] 2025-07-17T08:36:12.4061599Z + [[ -d .ci/docker ]] 2025-07-17T08:36:12.4061792Z + [[ -f .ci/docker/./build.sh ]] 2025-07-17T08:36:12.4061986Z + [[ true == \t\r\u\e ]] 2025-07-17T08:36:12.4062167Z + echo skip=false 2025-07-17T08:36:12.4063119Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-07-17T08:36:12.4072351Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:12.4072815Z ++ awk -F '[:,]' '{print $2}' 2025-07-17T08:36:12.4087929Z + DOCKER_TAG=pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:12.4088446Z + echo docker-tag=pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:12.4089085Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:12.4119419Z ##[group]Run set +e 2025-07-17T08:36:12.4119647Z set +e 2025-07-17T08:36:12.4119805Z set -x 2025-07-17T08:36:12.4119953Z  2025-07-17T08:36:12.4120108Z login() { 2025-07-17T08:36:12.4120424Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-07-17T08:36:12.4120760Z } 2025-07-17T08:36:12.4120905Z  2025-07-17T08:36:12.4121046Z retry () { 2025-07-17T08:36:12.4121221Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-07-17T08:36:12.4121426Z } 2025-07-17T08:36:12.4121564Z  2025-07-17T08:36:12.4121714Z retry login "${DOCKER_REGISTRY}" 2025-07-17T08:36:12.4121909Z  2025-07-17T08:36:12.4122055Z START_TIME=$(date +%s) 2025-07-17T08:36:12.4122468Z # Wait up to 120 minutes 2025-07-17T08:36:12.4122713Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-07-17T08:36:12.4123037Z  # Check if image already exists, if it does then skip building it 2025-07-17T08:36:12.4123341Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-07-17T08:36:12.4123575Z  exit 0 2025-07-17T08:36:12.4123735Z  fi 2025-07-17T08:36:12.4123889Z  2025-07-17T08:36:12.4124142Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-07-17T08:36:12.4124544Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-07-17T08:36:12.4124940Z  # latter, it will wait for the Docker images to become available before continuing 2025-07-17T08:36:12.4125262Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-07-17T08:36:12.4125523Z  # It's a Docker build job, let's build the image 2025-07-17T08:36:12.4125753Z  break 2025-07-17T08:36:12.4125912Z  else 2025-07-17T08:36:12.4126137Z  # It's a regular build job, wait for the image to become available 2025-07-17T08:36:12.4126397Z  sleep 300 2025-07-17T08:36:12.4126558Z  fi 2025-07-17T08:36:12.4126707Z done 2025-07-17T08:36:12.4126851Z  2025-07-17T08:36:12.4127078Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-07-17T08:36:12.4127603Z # be empty. The default action would be to continue rebuild the image 2025-07-17T08:36:12.4127910Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-07-17T08:36:12.4128189Z  # if we're on the base branch then use the parent commit 2025-07-17T08:36:12.4128440Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-07-17T08:36:12.4128776Z else 2025-07-17T08:36:12.4129083Z  # otherwise we're on a PR, so use the most recent base commit 2025-07-17T08:36:12.4129551Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-07-17T08:36:12.4129853Z fi 2025-07-17T08:36:12.4130108Z  2025-07-17T08:36:12.4130365Z if [[ -z "${MERGE_BASE}" ]]; then 2025-07-17T08:36:12.4130716Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-07-17T08:36:12.4131025Z  2025-07-17T08:36:12.4131438Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-07-17T08:36:12.4131886Z  exit 0 2025-07-17T08:36:12.4132124Z fi 2025-07-17T08:36:12.4132410Z  2025-07-17T08:36:12.4132706Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-07-17T08:36:12.4133216Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-07-17T08:36:12.4147436Z  exit 1 2025-07-17T08:36:12.4147627Z fi 2025-07-17T08:36:12.4147804Z  2025-07-17T08:36:12.4148059Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-07-17T08:36:12.4148490Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-07-17T08:36:12.4148863Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-07-17T08:36:12.4149279Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-07-17T08:36:12.4149744Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-07-17T08:36:12.4150030Z fi 2025-07-17T08:36:12.4150197Z  2025-07-17T08:36:12.4150380Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-07-17T08:36:12.4162645Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:12.4162897Z env: 2025-07-17T08:36:12.4163059Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:12.4163476Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:36:12.4163781Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:36:12.4164058Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:36:12.4164755Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:36:12.4165424Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:36:12.4165616Z AWS_REGION: us-east-1 2025-07-17T08:36:12.4165896Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:36:12.4166168Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:36:12.4169962Z AWS_SESSION_TOKEN: *** 2025-07-17T08:36:12.4170158Z DOCKER_BUILD_DIR: .ci/docker 2025-07-17T08:36:12.4170393Z BASE_REVISION: a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:36:12.4170916Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:12.4171498Z DOCKER_TAG: pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:12.4171867Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:12.4172119Z DOCKER_PUSH: 2025-07-17T08:36:12.4172277Z ##[endgroup] 2025-07-17T08:36:12.4201144Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:12.4201454Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:12.4204549Z + aws ecr get-login-password --region us-east-1 2025-07-17T08:36:12.4206005Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:12.4206452Z /home/runner/_work/_temp/ccaa283e-7d66-4f33-9f10-0a33eaaf54d9.sh: line 5: aws: command not found 2025-07-17T08:36:12.4357518Z Error: Cannot perform an interactive login from a non TTY device 2025-07-17T08:36:12.4377828Z + sleep 1 2025-07-17T08:36:13.4390835Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:13.4394183Z + aws ecr get-login-password --region us-east-1 2025-07-17T08:36:13.4395000Z /home/runner/_work/_temp/ccaa283e-7d66-4f33-9f10-0a33eaaf54d9.sh: line 5: aws: command not found 2025-07-17T08:36:13.4395499Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:13.4523918Z Error: Cannot perform an interactive login from a non TTY device 2025-07-17T08:36:13.4547260Z + sleep 2 2025-07-17T08:36:15.4562519Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:15.4566430Z + aws ecr get-login-password --region us-east-1 2025-07-17T08:36:15.4567025Z /home/runner/_work/_temp/ccaa283e-7d66-4f33-9f10-0a33eaaf54d9.sh: line 5: aws: command not found 2025-07-17T08:36:15.4567517Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:15.4725456Z Error: Cannot perform an interactive login from a non TTY device 2025-07-17T08:36:15.4752388Z ++ date +%s 2025-07-17T08:36:15.4762503Z + START_TIME=1752741375 2025-07-17T08:36:15.4767021Z ++ date +%s 2025-07-17T08:36:15.4776472Z + [[ 1752734175 -lt 1752741375 ]] 2025-07-17T08:36:15.4777036Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:16.0448088Z { 2025-07-17T08:36:16.0448411Z "schemaVersion": 2, 2025-07-17T08:36:16.0448862Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-07-17T08:36:16.0449197Z "config": { 2025-07-17T08:36:16.0449453Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-07-17T08:36:16.0449749Z "size": 28432, 2025-07-17T08:36:16.0450050Z "digest": "sha256:bf01532184142d9c9b40645b5f7203579796b76209b6d67c95931cea092c88ce" 2025-07-17T08:36:16.0450345Z }, 2025-07-17T08:36:16.0450488Z "layers": [ 2025-07-17T08:36:16.0450632Z { 2025-07-17T08:36:16.0451660Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0451969Z "size": 30590750, 2025-07-17T08:36:16.0452240Z "digest": "sha256:e87500e698966458d9dfc34df84602985c9821f39666619792fe6282aa6df5d4" 2025-07-17T08:36:16.0452523Z }, 2025-07-17T08:36:16.0452662Z { 2025-07-17T08:36:16.0452873Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0453130Z "size": 1554, 2025-07-17T08:36:16.0453402Z "digest": "sha256:3e261c1b63f9fafd9de479c9e09986e4995cb43ddf3dfc856377b165d15c917e" 2025-07-17T08:36:16.0453694Z }, 2025-07-17T08:36:16.0453826Z { 2025-07-17T08:36:16.0454033Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0454276Z "size": 336398964, 2025-07-17T08:36:16.0454543Z "digest": "sha256:6525f0fe3e6632053e95a1580897998453d74f9f2869cba6ad0e6df29ddaa00f" 2025-07-17T08:36:16.0454832Z }, 2025-07-17T08:36:16.0454968Z { 2025-07-17T08:36:16.0455176Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0455442Z "size": 703, 2025-07-17T08:36:16.0455708Z "digest": "sha256:52243ee284ea014545f5eb76a63e9556e2ee74a17ff461b08656265c3a82a05d" 2025-07-17T08:36:16.0455990Z }, 2025-07-17T08:36:16.0456124Z { 2025-07-17T08:36:16.0456339Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0456595Z "size": 1767, 2025-07-17T08:36:16.0456845Z "digest": "sha256:01fd0a5c43938350aeec33bb5fdd60818506892e41653d7cb81bf0d2df4aee49" 2025-07-17T08:36:16.0463479Z }, 2025-07-17T08:36:16.0463611Z { 2025-07-17T08:36:16.0463823Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0464078Z "size": 485, 2025-07-17T08:36:16.0464336Z "digest": "sha256:9fbeb7a0b287a2613a0b28c2559c25b6462554007dd34d36c1ea6a59eb54b0f1" 2025-07-17T08:36:16.0464625Z }, 2025-07-17T08:36:16.0464758Z { 2025-07-17T08:36:16.0464959Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0465219Z "size": 120654947, 2025-07-17T08:36:16.0465495Z "digest": "sha256:9a4035c74e1a9cff259d67cfedab4bf7b109688f88402c8682a389464941baa0" 2025-07-17T08:36:16.0465773Z }, 2025-07-17T08:36:16.0465914Z { 2025-07-17T08:36:16.0466115Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0466358Z "size": 4086, 2025-07-17T08:36:16.0466623Z "digest": "sha256:ea39ccd54b2768a871773082a1abf9fc4f1431146d23d9dfca53d062fc831d6c" 2025-07-17T08:36:16.0466913Z }, 2025-07-17T08:36:16.0467043Z { 2025-07-17T08:36:16.0467241Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0467486Z "size": 1709, 2025-07-17T08:36:16.0467735Z "digest": "sha256:73138d519c29056db37e950de740ce65ca83e9de7f4ba4712be32d4140c28a45" 2025-07-17T08:36:16.0468020Z }, 2025-07-17T08:36:16.0468140Z { 2025-07-17T08:36:16.0468351Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0468609Z "size": 724, 2025-07-17T08:36:16.0468870Z "digest": "sha256:26819d564d7ac5d2f9a72b83474ed2b2ded76d119361e40ce02e231bc477cd18" 2025-07-17T08:36:16.0469159Z }, 2025-07-17T08:36:16.0469297Z { 2025-07-17T08:36:16.0469518Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0469785Z "size": 3557009122, 2025-07-17T08:36:16.0470058Z "digest": "sha256:5f668547ed0d22644023a782e2bc34fdbb94d787bf79336e9bed2b2b4c67e9dd" 2025-07-17T08:36:16.0470350Z }, 2025-07-17T08:36:16.0470470Z { 2025-07-17T08:36:16.0470670Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0470917Z "size": 381, 2025-07-17T08:36:16.0471172Z "digest": "sha256:f8910737f31e7556508a27a8935eda8f3624b18e9e227a5a0055adbd305d4de7" 2025-07-17T08:36:16.0471450Z }, 2025-07-17T08:36:16.0471584Z { 2025-07-17T08:36:16.0471786Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0472039Z "size": 65128131, 2025-07-17T08:36:16.0472502Z "digest": "sha256:2548b702883d5a5fe514bd9c89f3549390855504fa205e91377a07c69b1a800a" 2025-07-17T08:36:16.0472799Z }, 2025-07-17T08:36:16.0472935Z { 2025-07-17T08:36:16.0473139Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0473404Z "size": 791, 2025-07-17T08:36:16.0474020Z "digest": "sha256:f949a24ea14864f5def0a329aee8040cb25534b81b6c3877fd6614e424d5ee91" 2025-07-17T08:36:16.0474346Z }, 2025-07-17T08:36:16.0474496Z { 2025-07-17T08:36:16.0474722Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0474987Z "size": 104, 2025-07-17T08:36:16.0475243Z "digest": "sha256:88bca81852d382f76fa109bbdfcb7a9745e9d00597998dd9c5a0895a1774eb09" 2025-07-17T08:36:16.0475550Z }, 2025-07-17T08:36:16.0475681Z { 2025-07-17T08:36:16.0475896Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0476147Z "size": 1495, 2025-07-17T08:36:16.0476410Z "digest": "sha256:9c1dbde76e68d2f5dfec6c94103995a25707f0a84dd34ba6ae99b15ceaa1b1e3" 2025-07-17T08:36:16.0476704Z }, 2025-07-17T08:36:16.0476838Z { 2025-07-17T08:36:16.0477041Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0477296Z "size": 541245038, 2025-07-17T08:36:16.0477565Z "digest": "sha256:1a45a3124fac16847c902c4acb5919cc6658a2e24c024e0f9822fdeb5ddb2c9d" 2025-07-17T08:36:16.0477843Z }, 2025-07-17T08:36:16.0477962Z { 2025-07-17T08:36:16.0478315Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0478564Z "size": 164, 2025-07-17T08:36:16.0478808Z "digest": "sha256:44222d017546d9b479e800c228767826ffed0d89439e900570a893d1b9f7f484" 2025-07-17T08:36:16.0479095Z }, 2025-07-17T08:36:16.0479235Z { 2025-07-17T08:36:16.0479453Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0479715Z "size": 2370, 2025-07-17T08:36:16.0479985Z "digest": "sha256:3de0bbb94b9029a81f9076c2ea6348035df964ba1d06cf7c3ca65d02daf0bdf9" 2025-07-17T08:36:16.0480281Z }, 2025-07-17T08:36:16.0480417Z { 2025-07-17T08:36:16.0480628Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0480880Z "size": 7163562192, 2025-07-17T08:36:16.0481161Z "digest": "sha256:396eca0560d4e2b2bbaa72276ec3a4bbfc94cb2ade43716f03b231ccc4877de6" 2025-07-17T08:36:16.0481452Z }, 2025-07-17T08:36:16.0481586Z { 2025-07-17T08:36:16.0481788Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0482043Z "size": 105, 2025-07-17T08:36:16.0482299Z "digest": "sha256:f7e0e33f00d762ece892a4880d65de4f60873bb2651a4e7adde882afc2f4c776" 2025-07-17T08:36:16.0482581Z }, 2025-07-17T08:36:16.0482713Z { 2025-07-17T08:36:16.0482903Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0483152Z "size": 612, 2025-07-17T08:36:16.0483397Z "digest": "sha256:ff78e5c67765635119ca736cdea8234ed1e53a62f934d6a71204491ec87a33f2" 2025-07-17T08:36:16.0483676Z }, 2025-07-17T08:36:16.0483805Z { 2025-07-17T08:36:16.0484020Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0484278Z "size": 677677280, 2025-07-17T08:36:16.0484551Z "digest": "sha256:103eb6a3b8850984365a0602f2cc9f6583d66be80ede8187fa109fa40f265bfd" 2025-07-17T08:36:16.0484836Z }, 2025-07-17T08:36:16.0484976Z { 2025-07-17T08:36:16.0485189Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0485455Z "size": 111, 2025-07-17T08:36:16.0485722Z "digest": "sha256:b28bda57e36a626d1ce025ad785984e424a7892a5c967c2be6c98982fa81d3a6" 2025-07-17T08:36:16.0486014Z }, 2025-07-17T08:36:16.0486149Z { 2025-07-17T08:36:16.0486363Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0486622Z "size": 1556, 2025-07-17T08:36:16.0486893Z "digest": "sha256:fd7bdb75a0223c4c6ba00af0bbfceca64a9584d0778cf1e301eda46a9fa2f753" 2025-07-17T08:36:16.0487183Z }, 2025-07-17T08:36:16.0487317Z { 2025-07-17T08:36:16.0487669Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0487922Z "size": 107, 2025-07-17T08:36:16.0488161Z "digest": "sha256:2baf05e68c264c8b5eb8c5954a8e8d1ae153e1ae935cc220c316e36746a82432" 2025-07-17T08:36:16.0488446Z }, 2025-07-17T08:36:16.0488580Z { 2025-07-17T08:36:16.0488777Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0489084Z + exit 0 2025-07-17T08:36:16.0489223Z "size": 166, 2025-07-17T08:36:16.0489477Z "digest": "sha256:c8d437fa52e4cdd7d36b0bed532b2cf938fd68314d18dcdf3d5cf55ef63e5e1d" 2025-07-17T08:36:16.0489768Z }, 2025-07-17T08:36:16.0489896Z { 2025-07-17T08:36:16.0490095Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0490343Z "size": 2894072, 2025-07-17T08:36:16.0490597Z "digest": "sha256:e7f0c42a41506fb32e80839aeef946a1c06bd2670ad817031fa69259064ee4e4" 2025-07-17T08:36:16.0490877Z }, 2025-07-17T08:36:16.0491005Z { 2025-07-17T08:36:16.0491212Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0491457Z "size": 107, 2025-07-17T08:36:16.0491691Z "digest": "sha256:0459eb37c565555487c7fec1289ce0e5a5f353aa0a5c6601262cd727c696db6e" 2025-07-17T08:36:16.0491972Z }, 2025-07-17T08:36:16.0492103Z { 2025-07-17T08:36:16.0492304Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0492552Z "size": 802, 2025-07-17T08:36:16.0492921Z "digest": "sha256:54bf6ee2f94ce79d68059340d3ed4a1efbbe4813b8ace577b80539c45c947791" 2025-07-17T08:36:16.0493202Z }, 2025-07-17T08:36:16.0493333Z { 2025-07-17T08:36:16.0493529Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0493786Z "size": 26116470, 2025-07-17T08:36:16.0494053Z "digest": "sha256:7e4f3e71a58d6a44e3c9a978d14a30dca6c3cc8e4fb155b802db9a9a6f619794" 2025-07-17T08:36:16.0494339Z }, 2025-07-17T08:36:16.0494472Z { 2025-07-17T08:36:16.0494669Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0494925Z "size": 104, 2025-07-17T08:36:16.0495178Z "digest": "sha256:11d68569165e3d3f65fd7d665a068100e7cded4ae1d9b296d9103cd8f267af99" 2025-07-17T08:36:16.0495458Z }, 2025-07-17T08:36:16.0495593Z { 2025-07-17T08:36:16.0495794Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0496038Z "size": 425, 2025-07-17T08:36:16.0496283Z "digest": "sha256:703d327e6498dc9823aac0404c90e832a14b922d1671fd8c4cf0c484a192bd2f" 2025-07-17T08:36:16.0496564Z }, 2025-07-17T08:36:16.0496685Z { 2025-07-17T08:36:16.0496884Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0497151Z "size": 19316525, 2025-07-17T08:36:16.0497410Z "digest": "sha256:e73111c3b4ee3754e1110154ede278434b332e63f88403fd8be1eb25791556d3" 2025-07-17T08:36:16.0497687Z }, 2025-07-17T08:36:16.0497815Z { 2025-07-17T08:36:16.0498015Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0498257Z "size": 425, 2025-07-17T08:36:16.0498512Z "digest": "sha256:de0a219b38cf6c78fb6c14f71dd0262d917740046f8b3ac5ee29621b318dbf4d" 2025-07-17T08:36:16.0498802Z }, 2025-07-17T08:36:16.0498937Z { 2025-07-17T08:36:16.0499128Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0499377Z "size": 724, 2025-07-17T08:36:16.0499637Z "digest": "sha256:26819d564d7ac5d2f9a72b83474ed2b2ded76d119361e40ce02e231bc477cd18" 2025-07-17T08:36:16.0499924Z }, 2025-07-17T08:36:16.0500062Z { 2025-07-17T08:36:16.0500270Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0500528Z "size": 141, 2025-07-17T08:36:16.0500777Z "digest": "sha256:ff5883ff248637e8a5e9c52d7dfd140bd63da71d63052c46743bd754884a8edb" 2025-07-17T08:36:16.0501057Z }, 2025-07-17T08:36:16.0501188Z { 2025-07-17T08:36:16.0501380Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0501629Z "size": 137, 2025-07-17T08:36:16.0502029Z "digest": "sha256:2cb94b4ea22d39df80711c10a039e66f91df4ce7df7cc77d3bf87dc0813b5415" 2025-07-17T08:36:16.0502321Z }, 2025-07-17T08:36:16.0502458Z { 2025-07-17T08:36:16.0502660Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0502909Z "size": 5103087710, 2025-07-17T08:36:16.0503172Z "digest": "sha256:85520474c47bbf53e96952e3c156b595e84eb16762dc6413140cb018c296c04a" 2025-07-17T08:36:16.0503448Z }, 2025-07-17T08:36:16.0503574Z { 2025-07-17T08:36:16.0503770Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0504008Z "size": 194, 2025-07-17T08:36:16.0504253Z "digest": "sha256:4995dd45751be27343cb11b07f035f1c1e1973be4ce5cb8a9d92dc16e1f5c453" 2025-07-17T08:36:16.0504534Z }, 2025-07-17T08:36:16.0504666Z { 2025-07-17T08:36:16.0504867Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0505119Z "size": 347, 2025-07-17T08:36:16.0505380Z "digest": "sha256:05db2817f2540d1802df97148a259d0d59ceadbc6aedcdaaca8ab62f55171bac" 2025-07-17T08:36:16.0505667Z }, 2025-07-17T08:36:16.0505801Z { 2025-07-17T08:36:16.0506000Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0506252Z "size": 88297, 2025-07-17T08:36:16.0506503Z "digest": "sha256:07e5f9cd2b10db21ec946c38e2d1c768d1beb014a857a08dcceba16ca1ddf35d" 2025-07-17T08:36:16.0506789Z }, 2025-07-17T08:36:16.0507050Z { 2025-07-17T08:36:16.0507258Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0507507Z "size": 106, 2025-07-17T08:36:16.0507764Z "digest": "sha256:21dab22df0ecba699c9df688c9d1da5d373f1856b7756886d78695881f1df8ee" 2025-07-17T08:36:16.0508046Z }, 2025-07-17T08:36:16.0508182Z { 2025-07-17T08:36:16.0508380Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0508625Z "size": 1666, 2025-07-17T08:36:16.0508875Z "digest": "sha256:cd89850e2c442e231bf3107d7c9ebf10cefa69feb439e39eaf0224cf943df949" 2025-07-17T08:36:16.0509153Z }, 2025-07-17T08:36:16.0509284Z { 2025-07-17T08:36:16.0509484Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0509740Z "size": 724, 2025-07-17T08:36:16.0510001Z "digest": "sha256:26819d564d7ac5d2f9a72b83474ed2b2ded76d119361e40ce02e231bc477cd18" 2025-07-17T08:36:16.0510289Z }, 2025-07-17T08:36:16.0510430Z { 2025-07-17T08:36:16.0510650Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0510902Z "size": 137, 2025-07-17T08:36:16.0511158Z "digest": "sha256:7290f4ad3db1f8a188a62bd95fb699be55cc0f928d45b6878c2aefc972e8cf42" 2025-07-17T08:36:16.0511449Z }, 2025-07-17T08:36:16.0511575Z { 2025-07-17T08:36:16.0511779Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0512030Z "size": 120, 2025-07-17T08:36:16.0512286Z "digest": "sha256:ee114be1a250f21b8e9574d1fcf4e6fe619b9fcc60d91849514875fe29659e86" 2025-07-17T08:36:16.0512567Z }, 2025-07-17T08:36:16.0512713Z { 2025-07-17T08:36:16.0512920Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0513186Z "size": 5337400283, 2025-07-17T08:36:16.0513455Z "digest": "sha256:88d77fb451a4cbb5d7c86b95cc850f83679bb2f75fc7a931023cf8166e4c30e7" 2025-07-17T08:36:16.0513742Z }, 2025-07-17T08:36:16.0513937Z { 2025-07-17T08:36:16.0514150Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0514412Z "size": 175, 2025-07-17T08:36:16.0514682Z "digest": "sha256:eb98dfa00416bfd6b6d30396e39b46cb2d85205a25108022278c641176a1023a" 2025-07-17T08:36:16.0514965Z }, 2025-07-17T08:36:16.0515093Z { 2025-07-17T08:36:16.0515291Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0515539Z "size": 1897, 2025-07-17T08:36:16.0515784Z "digest": "sha256:0971ac0664a4884417ddb7be991372a3420280874578d17dc678b3d2ef293c3a" 2025-07-17T08:36:16.0516061Z }, 2025-07-17T08:36:16.0516203Z { 2025-07-17T08:36:16.0516584Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0516834Z "size": 162687915, 2025-07-17T08:36:16.0517099Z "digest": "sha256:9b9239f68e90235feae6f403c501599dea179aca869918e48adbe25ddfac9bf4" 2025-07-17T08:36:16.0517378Z }, 2025-07-17T08:36:16.0517503Z { 2025-07-17T08:36:16.0517707Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0517967Z "size": 304, 2025-07-17T08:36:16.0518218Z "digest": "sha256:4d3441d455038ca68278ebfb4461fac6943dfe5632034e6dcc202f1455ea1842" 2025-07-17T08:36:16.0518496Z }, 2025-07-17T08:36:16.0518627Z { 2025-07-17T08:36:16.0518819Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0519060Z "size": 32, 2025-07-17T08:36:16.0519314Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-07-17T08:36:16.0519594Z }, 2025-07-17T08:36:16.0519730Z { 2025-07-17T08:36:16.0519930Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0520178Z "size": 108, 2025-07-17T08:36:16.0520429Z "digest": "sha256:bc1ff02b1df01038762aecf59ce5f90041fe38b68e9b53988bac836a26398de3" 2025-07-17T08:36:16.0520711Z }, 2025-07-17T08:36:16.0520847Z { 2025-07-17T08:36:16.0521044Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-07-17T08:36:16.0521295Z "size": 54145699, 2025-07-17T08:36:16.0521709Z "digest": "sha256:e20c94532338189b44d993e4075ee1bda5c2e71da7f8e3ec6812b96dbc85502e" 2025-07-17T08:36:16.0522006Z } 2025-07-17T08:36:16.0522131Z ] 2025-07-17T08:36:16.0522258Z } 2025-07-17T08:36:16.0551771Z ##[group]Run set -eux 2025-07-17T08:36:16.0551983Z set -eux 2025-07-17T08:36:16.0552525Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin 2025-07-17T08:36:16.0565279Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:16.0565533Z env: 2025-07-17T08:36:16.0565688Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:16.0565907Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:36:16.0566197Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:36:16.0566467Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:36:16.0567156Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:36:16.0567819Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:36:16.0568009Z AWS_REGION: us-east-1 2025-07-17T08:36:16.0568298Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:36:16.0568554Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:36:16.0572331Z AWS_SESSION_TOKEN: *** 2025-07-17T08:36:16.0572510Z ##[endgroup] 2025-07-17T08:36:16.0611759Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-07-17T08:36:16.0612239Z /home/runner/_work/_temp/dc6a69fd-742d-490e-bc3c-599cfda678ab.sh: line 2: aws: command not found 2025-07-17T08:36:16.0612599Z + jq --raw-output .SecretString 2025-07-17T08:36:16.0613338Z + jq -r .docker_hub_readonly_token 2025-07-17T08:36:16.0614623Z + docker login --username pytorchbot --password-stdin 2025-07-17T08:36:16.0774868Z Error: Cannot perform an interactive login from a non TTY device 2025-07-17T08:36:16.0802802Z ##[error]Process completed with exit code 1. 2025-07-17T08:36:16.0900647Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-07-17T08:36:16.0900939Z with: 2025-07-17T08:36:16.0901388Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:16.0901926Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:16.0902185Z env: 2025-07-17T08:36:16.0902352Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:16.0902585Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:36:16.0902888Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:36:16.0903168Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:36:16.0903914Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:36:16.0904592Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:36:16.0904788Z AWS_REGION: us-east-1 2025-07-17T08:36:16.0905077Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:36:16.0905337Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:36:16.0909149Z AWS_SESSION_TOKEN: *** 2025-07-17T08:36:16.0909335Z ##[endgroup] 2025-07-17T08:36:16.0924304Z ##[group]Run set -x 2025-07-17T08:36:16.0924531Z set -x 2025-07-17T08:36:16.0924693Z set +e 2025-07-17T08:36:16.0924841Z  2025-07-17T08:36:16.0924994Z login() { 2025-07-17T08:36:16.0925326Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-07-17T08:36:16.0925660Z } 2025-07-17T08:36:16.0925811Z  2025-07-17T08:36:16.0925959Z retry () { 2025-07-17T08:36:16.0926423Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-07-17T08:36:16.0926644Z } 2025-07-17T08:36:16.0926784Z  2025-07-17T08:36:16.0926947Z retry login "${DOCKER_REGISTRY}" 2025-07-17T08:36:16.0927153Z  2025-07-17T08:36:16.0927471Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-07-17T08:36:16.0927877Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-07-17T08:36:16.0928124Z  2025-07-17T08:36:16.0928278Z set -e 2025-07-17T08:36:16.0928502Z # ignore output since only exit code is used for conditional 2025-07-17T08:36:16.0928810Z # only pull docker image if it's not available locally 2025-07-17T08:36:16.0929150Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-07-17T08:36:16.0929471Z  retry docker pull "${DOCKER_IMAGE}" 2025-07-17T08:36:16.0929671Z fi 2025-07-17T08:36:16.0941312Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:36:16.0941555Z env: 2025-07-17T08:36:16.0941711Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:36:16.0941943Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:36:16.0942234Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:36:16.0942507Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:36:16.0943195Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:36:16.0943855Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:36:16.0944043Z AWS_REGION: us-east-1 2025-07-17T08:36:16.0944274Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:36:16.0944539Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:36:16.0948351Z AWS_SESSION_TOKEN: *** 2025-07-17T08:36:16.0949047Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:16.0949581Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:16.0949834Z ##[endgroup] 2025-07-17T08:36:16.0998958Z + set +e 2025-07-17T08:36:16.0999264Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:16.0999578Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:16.1002200Z + aws ecr get-login-password --region us-east-1 2025-07-17T08:36:16.1002592Z /home/runner/_work/_temp/c056e681-4283-4b06-9a13-96fb0e67ce79.sh: line 5: aws: command not found 2025-07-17T08:36:16.1003431Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:16.1166607Z Error: Cannot perform an interactive login from a non TTY device 2025-07-17T08:36:16.1182691Z + sleep 1 2025-07-17T08:36:17.1195787Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:17.1199922Z + aws ecr get-login-password --region us-east-1 2025-07-17T08:36:17.1200354Z /home/runner/_work/_temp/c056e681-4283-4b06-9a13-96fb0e67ce79.sh: line 5: aws: command not found 2025-07-17T08:36:17.1200805Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:17.1352971Z Error: Cannot perform an interactive login from a non TTY device 2025-07-17T08:36:17.1374253Z + sleep 2 2025-07-17T08:36:19.1389413Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:19.1393437Z + aws ecr get-login-password --region us-east-1 2025-07-17T08:36:19.1394661Z /home/runner/_work/_temp/c056e681-4283-4b06-9a13-96fb0e67ce79.sh: line 5: aws: command not found 2025-07-17T08:36:19.1395168Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T08:36:19.1543602Z Error: Cannot perform an interactive login from a non TTY device 2025-07-17T08:36:19.1569909Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:19.1571139Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-07-17T08:36:19.8057724Z + IMAGE_SIZE=22123.39352798462 2025-07-17T08:36:19.8058016Z Compressed size of image in MB: 22123.39352798462 2025-07-17T08:36:19.8058305Z + echo 'Compressed size of image in MB: 22123.39352798462' 2025-07-17T08:36:19.8058562Z + set -e 2025-07-17T08:36:19.8059037Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:19.8265906Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:19.8266711Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:36:20.3606808Z pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d: Pulling from pytorch/ci-image 2025-07-17T08:36:20.3614218Z e87500e69896: Pulling fs layer 2025-07-17T08:36:20.3614457Z 3e261c1b63f9: Pulling fs layer 2025-07-17T08:36:20.3614779Z 6525f0fe3e66: Pulling fs layer 2025-07-17T08:36:20.3615024Z 52243ee284ea: Pulling fs layer 2025-07-17T08:36:20.3615216Z 01fd0a5c4393: Pulling fs layer 2025-07-17T08:36:20.3615393Z 9fbeb7a0b287: Pulling fs layer 2025-07-17T08:36:20.3615570Z 9a4035c74e1a: Pulling fs layer 2025-07-17T08:36:20.3615755Z ea39ccd54b27: Pulling fs layer 2025-07-17T08:36:20.3615938Z 73138d519c29: Pulling fs layer 2025-07-17T08:36:20.3616109Z 26819d564d7a: Pulling fs layer 2025-07-17T08:36:20.3616285Z 5f668547ed0d: Pulling fs layer 2025-07-17T08:36:20.3616455Z f8910737f31e: Pulling fs layer 2025-07-17T08:36:20.3616624Z 2548b702883d: Pulling fs layer 2025-07-17T08:36:20.3616794Z f949a24ea148: Pulling fs layer 2025-07-17T08:36:20.3616965Z 88bca81852d3: Pulling fs layer 2025-07-17T08:36:20.3617156Z 9c1dbde76e68: Pulling fs layer 2025-07-17T08:36:20.3617336Z 1a45a3124fac: Pulling fs layer 2025-07-17T08:36:20.3618088Z 44222d017546: Pulling fs layer 2025-07-17T08:36:20.3618258Z 3de0bbb94b90: Pulling fs layer 2025-07-17T08:36:20.3618431Z 396eca0560d4: Pulling fs layer 2025-07-17T08:36:20.3618596Z 52243ee284ea: Waiting 2025-07-17T08:36:20.3618770Z f7e0e33f00d7: Pulling fs layer 2025-07-17T08:36:20.3618952Z ff78e5c67765: Pulling fs layer 2025-07-17T08:36:20.3619168Z 103eb6a3b885: Pulling fs layer 2025-07-17T08:36:20.3619338Z b28bda57e36a: Pulling fs layer 2025-07-17T08:36:20.3619519Z fd7bdb75a022: Pulling fs layer 2025-07-17T08:36:20.3619692Z 2baf05e68c26: Pulling fs layer 2025-07-17T08:36:20.3619858Z 26819d564d7a: Waiting 2025-07-17T08:36:20.3620017Z c8d437fa52e4: Pulling fs layer 2025-07-17T08:36:20.3620197Z f949a24ea148: Waiting 2025-07-17T08:36:20.3620352Z 5f668547ed0d: Waiting 2025-07-17T08:36:20.3620493Z 01fd0a5c4393: Waiting 2025-07-17T08:36:20.3620651Z e7f0c42a4150: Pulling fs layer 2025-07-17T08:36:20.3620824Z f8910737f31e: Waiting 2025-07-17T08:36:20.3620980Z 88bca81852d3: Waiting 2025-07-17T08:36:20.3621133Z 9a4035c74e1a: Waiting 2025-07-17T08:36:20.3621281Z ea39ccd54b27: Waiting 2025-07-17T08:36:20.3621438Z 0459eb37c565: Pulling fs layer 2025-07-17T08:36:20.3621610Z 3de0bbb94b90: Waiting 2025-07-17T08:36:20.3621764Z 9fbeb7a0b287: Waiting 2025-07-17T08:36:20.3621918Z f7e0e33f00d7: Waiting 2025-07-17T08:36:20.3622059Z 73138d519c29: Waiting 2025-07-17T08:36:20.3622217Z 54bf6ee2f94c: Pulling fs layer 2025-07-17T08:36:20.3622395Z 7e4f3e71a58d: Pulling fs layer 2025-07-17T08:36:20.3622563Z ff78e5c67765: Waiting 2025-07-17T08:36:20.3622715Z 11d68569165e: Pulling fs layer 2025-07-17T08:36:20.3622883Z 703d327e6498: Pulling fs layer 2025-07-17T08:36:20.3623052Z e73111c3b4ee: Pulling fs layer 2025-07-17T08:36:20.3623217Z 44222d017546: Waiting 2025-07-17T08:36:20.3623371Z de0a219b38cf: Pulling fs layer 2025-07-17T08:36:20.3623532Z 2548b702883d: Waiting 2025-07-17T08:36:20.3623871Z ff5883ff2486: Pulling fs layer 2025-07-17T08:36:20.3624041Z 103eb6a3b885: Waiting 2025-07-17T08:36:20.3624207Z 2cb94b4ea22d: Pulling fs layer 2025-07-17T08:36:20.3624394Z 85520474c47b: Pulling fs layer 2025-07-17T08:36:20.3624566Z 4995dd45751b: Pulling fs layer 2025-07-17T08:36:20.3624732Z b28bda57e36a: Waiting 2025-07-17T08:36:20.3624885Z 1a45a3124fac: Waiting 2025-07-17T08:36:20.3625040Z 05db2817f254: Pulling fs layer 2025-07-17T08:36:20.3625207Z fd7bdb75a022: Waiting 2025-07-17T08:36:20.3625357Z 2baf05e68c26: Waiting 2025-07-17T08:36:20.3625508Z 07e5f9cd2b10: Pulling fs layer 2025-07-17T08:36:20.3625679Z 21dab22df0ec: Pulling fs layer 2025-07-17T08:36:20.3625849Z cd89850e2c44: Pulling fs layer 2025-07-17T08:36:20.3626015Z c8d437fa52e4: Waiting 2025-07-17T08:36:20.3626165Z 7290f4ad3db1: Pulling fs layer 2025-07-17T08:36:20.3626334Z ee114be1a250: Pulling fs layer 2025-07-17T08:36:20.3626501Z 54bf6ee2f94c: Waiting 2025-07-17T08:36:20.3626655Z 88d77fb451a4: Pulling fs layer 2025-07-17T08:36:20.3626833Z eb98dfa00416: Pulling fs layer 2025-07-17T08:36:20.3627000Z 0971ac0664a4: Pulling fs layer 2025-07-17T08:36:20.3627160Z 7290f4ad3db1: Waiting 2025-07-17T08:36:20.3627307Z ee114be1a250: Waiting 2025-07-17T08:36:20.3627450Z 0459eb37c565: Waiting 2025-07-17T08:36:20.3627596Z 7e4f3e71a58d: Waiting 2025-07-17T08:36:20.3627747Z 9b9239f68e90: Pulling fs layer 2025-07-17T08:36:20.3627915Z 4d3441d45503: Pulling fs layer 2025-07-17T08:36:20.3628083Z 4f4fb700ef54: Pulling fs layer 2025-07-17T08:36:20.3628255Z bc1ff02b1df0: Pulling fs layer 2025-07-17T08:36:20.3628419Z de0a219b38cf: Waiting 2025-07-17T08:36:20.3628571Z e20c94532338: Pulling fs layer 2025-07-17T08:36:20.3628732Z cd89850e2c44: Waiting 2025-07-17T08:36:20.3628873Z 21dab22df0ec: Waiting 2025-07-17T08:36:20.3629021Z 88d77fb451a4: Waiting 2025-07-17T08:36:20.3629166Z 9b9239f68e90: Waiting 2025-07-17T08:36:20.3629311Z 0971ac0664a4: Waiting 2025-07-17T08:36:20.3629457Z 2cb94b4ea22d: Waiting 2025-07-17T08:36:20.3629603Z 05db2817f254: Waiting 2025-07-17T08:36:20.3629752Z 4f4fb700ef54: Waiting 2025-07-17T08:36:20.3629895Z 4d3441d45503: Waiting 2025-07-17T08:36:20.3630036Z 703d327e6498: Waiting 2025-07-17T08:36:20.3630300Z e73111c3b4ee: Waiting 2025-07-17T08:36:20.3630441Z 85520474c47b: Waiting 2025-07-17T08:36:20.3630587Z 11d68569165e: Waiting 2025-07-17T08:36:20.3630740Z eb98dfa00416: Waiting 2025-07-17T08:36:20.3630887Z bc1ff02b1df0: Waiting 2025-07-17T08:36:20.3631035Z e20c94532338: Waiting 2025-07-17T08:36:20.6128663Z 3e261c1b63f9: Verifying Checksum 2025-07-17T08:36:20.6128940Z 3e261c1b63f9: Download complete 2025-07-17T08:36:20.8618165Z 52243ee284ea: Download complete 2025-07-17T08:36:20.8807348Z e87500e69896: Verifying Checksum 2025-07-17T08:36:20.8807581Z e87500e69896: Download complete 2025-07-17T08:36:21.3054000Z 9fbeb7a0b287: Verifying Checksum 2025-07-17T08:36:21.3054287Z 9fbeb7a0b287: Download complete 2025-07-17T08:36:21.3070815Z 01fd0a5c4393: Verifying Checksum 2025-07-17T08:36:21.3071054Z 01fd0a5c4393: Download complete 2025-07-17T08:36:21.5549615Z ea39ccd54b27: Verifying Checksum 2025-07-17T08:36:21.5549879Z ea39ccd54b27: Download complete 2025-07-17T08:36:21.7967000Z 73138d519c29: Verifying Checksum 2025-07-17T08:36:21.7967301Z 73138d519c29: Download complete 2025-07-17T08:36:22.0322663Z 26819d564d7a: Verifying Checksum 2025-07-17T08:36:22.0322938Z 26819d564d7a: Download complete 2025-07-17T08:36:22.0804335Z e87500e69896: Pull complete 2025-07-17T08:36:22.2731082Z 3e261c1b63f9: Pull complete 2025-07-17T08:36:22.7450091Z 9a4035c74e1a: Verifying Checksum 2025-07-17T08:36:22.7450373Z 9a4035c74e1a: Download complete 2025-07-17T08:36:22.9759693Z f8910737f31e: Verifying Checksum 2025-07-17T08:36:22.9759981Z f8910737f31e: Download complete 2025-07-17T08:36:23.8494363Z 2548b702883d: Verifying Checksum 2025-07-17T08:36:23.8494647Z 2548b702883d: Download complete 2025-07-17T08:36:23.9326414Z 6525f0fe3e66: Verifying Checksum 2025-07-17T08:36:23.9326701Z 6525f0fe3e66: Download complete 2025-07-17T08:36:24.1017078Z f949a24ea148: Download complete 2025-07-17T08:36:24.1796728Z 88bca81852d3: Verifying Checksum 2025-07-17T08:36:24.1797007Z 88bca81852d3: Download complete 2025-07-17T08:36:24.3500365Z 9c1dbde76e68: Verifying Checksum 2025-07-17T08:36:24.3500654Z 9c1dbde76e68: Download complete 2025-07-17T08:36:24.5977440Z 44222d017546: Verifying Checksum 2025-07-17T08:36:24.5977708Z 44222d017546: Download complete 2025-07-17T08:36:24.8538059Z 3de0bbb94b90: Verifying Checksum 2025-07-17T08:36:24.8538334Z 3de0bbb94b90: Download complete 2025-07-17T08:36:29.7839362Z 1a45a3124fac: Verifying Checksum 2025-07-17T08:36:29.7839647Z 1a45a3124fac: Download complete 2025-07-17T08:36:30.0409952Z f7e0e33f00d7: Verifying Checksum 2025-07-17T08:36:30.2766218Z f7e0e33f00d7: Download complete 2025-07-17T08:36:30.2766488Z ff78e5c67765: Verifying Checksum 2025-07-17T08:36:30.2766695Z ff78e5c67765: Download complete 2025-07-17T08:36:32.7616833Z 6525f0fe3e66: Pull complete 2025-07-17T08:36:32.9498108Z 52243ee284ea: Pull complete 2025-07-17T08:36:33.1131022Z 01fd0a5c4393: Pull complete 2025-07-17T08:36:33.2668453Z 9fbeb7a0b287: Pull complete 2025-07-17T08:36:35.5196293Z 9a4035c74e1a: Pull complete 2025-07-17T08:36:35.7173662Z ea39ccd54b27: Pull complete 2025-07-17T08:36:35.9062949Z 73138d519c29: Pull complete 2025-07-17T08:36:36.0531359Z 26819d564d7a: Pull complete 2025-07-17T08:36:37.2853421Z 103eb6a3b885: Verifying Checksum 2025-07-17T08:36:37.2853696Z 103eb6a3b885: Download complete 2025-07-17T08:36:37.5300048Z b28bda57e36a: Verifying Checksum 2025-07-17T08:36:37.5300339Z b28bda57e36a: Download complete 2025-07-17T08:36:37.7732324Z fd7bdb75a022: Verifying Checksum 2025-07-17T08:36:37.7732615Z fd7bdb75a022: Download complete 2025-07-17T08:36:37.9999886Z 2baf05e68c26: Verifying Checksum 2025-07-17T08:36:38.0000261Z 2baf05e68c26: Download complete 2025-07-17T08:36:38.2288626Z c8d437fa52e4: Verifying Checksum 2025-07-17T08:36:38.2289020Z c8d437fa52e4: Download complete 2025-07-17T08:36:38.6448547Z e7f0c42a4150: Verifying Checksum 2025-07-17T08:36:38.6448831Z e7f0c42a4150: Download complete 2025-07-17T08:36:38.8710536Z 0459eb37c565: Verifying Checksum 2025-07-17T08:36:38.8710864Z 0459eb37c565: Download complete 2025-07-17T08:36:39.0829137Z 54bf6ee2f94c: Verifying Checksum 2025-07-17T08:36:39.0829444Z 54bf6ee2f94c: Download complete 2025-07-17T08:36:39.6516001Z 7e4f3e71a58d: Verifying Checksum 2025-07-17T08:36:39.6516282Z 7e4f3e71a58d: Download complete 2025-07-17T08:36:39.8882605Z 11d68569165e: Verifying Checksum 2025-07-17T08:36:39.8882878Z 11d68569165e: Download complete 2025-07-17T08:36:40.1238888Z 703d327e6498: Verifying Checksum 2025-07-17T08:36:40.1239176Z 703d327e6498: Download complete 2025-07-17T08:36:40.6432365Z e73111c3b4ee: Verifying Checksum 2025-07-17T08:36:40.6432698Z e73111c3b4ee: Download complete 2025-07-17T08:36:40.8993226Z de0a219b38cf: Verifying Checksum 2025-07-17T08:36:40.8993525Z de0a219b38cf: Download complete 2025-07-17T08:36:41.1312030Z ff5883ff2486: Verifying Checksum 2025-07-17T08:36:41.1312323Z ff5883ff2486: Download complete 2025-07-17T08:36:41.3782594Z 2cb94b4ea22d: Verifying Checksum 2025-07-17T08:36:41.3782952Z 2cb94b4ea22d: Download complete 2025-07-17T08:36:57.8309351Z 5f668547ed0d: Verifying Checksum 2025-07-17T08:36:57.8309717Z 5f668547ed0d: Download complete 2025-07-17T08:37:01.1996006Z 4995dd45751b: Verifying Checksum 2025-07-17T08:37:01.1996312Z 4995dd45751b: Download complete 2025-07-17T08:37:01.4390322Z 05db2817f254: Verifying Checksum 2025-07-17T08:37:01.4390767Z 05db2817f254: Download complete 2025-07-17T08:37:01.7163151Z 07e5f9cd2b10: Verifying Checksum 2025-07-17T08:37:01.7163470Z 07e5f9cd2b10: Download complete 2025-07-17T08:37:01.9389348Z 21dab22df0ec: Download complete 2025-07-17T08:37:02.1960430Z cd89850e2c44: Verifying Checksum 2025-07-17T08:37:02.1961679Z cd89850e2c44: Download complete 2025-07-17T08:37:02.4717140Z 7290f4ad3db1: Verifying Checksum 2025-07-17T08:37:02.4717530Z 7290f4ad3db1: Download complete 2025-07-17T08:37:02.7127445Z ee114be1a250: Verifying Checksum 2025-07-17T08:37:02.7127745Z ee114be1a250: Download complete 2025-07-17T08:37:46.1999577Z 5f668547ed0d: Pull complete 2025-07-17T08:37:46.5500323Z f8910737f31e: Pull complete 2025-07-17T08:37:47.6007360Z 2548b702883d: Pull complete 2025-07-17T08:37:47.8425254Z f949a24ea148: Pull complete 2025-07-17T08:37:48.0153680Z 88bca81852d3: Pull complete 2025-07-17T08:37:48.1808423Z 9c1dbde76e68: Pull complete 2025-07-17T08:37:54.9136196Z 1a45a3124fac: Pull complete 2025-07-17T08:37:55.1045382Z 44222d017546: Pull complete 2025-07-17T08:37:55.2580944Z 3de0bbb94b90: Pull complete 2025-07-17T08:38:41.6335637Z 396eca0560d4: Verifying Checksum 2025-07-17T08:38:41.6335948Z 396eca0560d4: Download complete 2025-07-17T08:38:42.1214762Z eb98dfa00416: Verifying Checksum 2025-07-17T08:38:42.1215091Z eb98dfa00416: Download complete 2025-07-17T08:38:42.3454066Z 0971ac0664a4: Verifying Checksum 2025-07-17T08:38:42.3454402Z 0971ac0664a4: Download complete 2025-07-17T08:38:44.2020791Z 9b9239f68e90: Verifying Checksum 2025-07-17T08:38:44.2021148Z 9b9239f68e90: Download complete 2025-07-17T08:38:44.4388235Z 4d3441d45503: Verifying Checksum 2025-07-17T08:38:44.4388589Z 4d3441d45503: Download complete 2025-07-17T08:38:44.5475305Z 4f4fb700ef54: Verifying Checksum 2025-07-17T08:38:44.5475697Z 4f4fb700ef54: Download complete 2025-07-17T08:38:44.7968270Z bc1ff02b1df0: Download complete 2025-07-17T08:38:45.5770420Z e20c94532338: Verifying Checksum 2025-07-17T08:38:55.8283069Z 88d77fb451a4: Download complete 2025-07-17T08:39:00.7519647Z 85520474c47b: Verifying Checksum 2025-07-17T08:39:00.7519923Z 85520474c47b: Download complete 2025-07-17T08:39:55.3934193Z 396eca0560d4: Pull complete 2025-07-17T08:39:56.1107904Z f7e0e33f00d7: Pull complete 2025-07-17T08:39:56.2454269Z ff78e5c67765: Pull complete 2025-07-17T08:40:02.2856260Z 103eb6a3b885: Pull complete 2025-07-17T08:40:02.4669739Z b28bda57e36a: Pull complete 2025-07-17T08:40:02.6103940Z fd7bdb75a022: Pull complete 2025-07-17T08:40:02.7539704Z 2baf05e68c26: Pull complete 2025-07-17T08:40:02.8997727Z c8d437fa52e4: Pull complete 2025-07-17T08:40:03.0685056Z e7f0c42a4150: Pull complete 2025-07-17T08:40:03.2054617Z 0459eb37c565: Pull complete 2025-07-17T08:40:03.3604554Z 54bf6ee2f94c: Pull complete 2025-07-17T08:40:03.8040685Z 7e4f3e71a58d: Pull complete 2025-07-17T08:40:03.9482396Z 11d68569165e: Pull complete 2025-07-17T08:40:04.0819912Z 703d327e6498: Pull complete 2025-07-17T08:40:04.4294065Z e73111c3b4ee: Pull complete 2025-07-17T08:40:04.5708393Z de0a219b38cf: Pull complete 2025-07-17T08:40:04.8627041Z ff5883ff2486: Pull complete 2025-07-17T08:40:05.0004217Z 2cb94b4ea22d: Pull complete 2025-07-17T08:40:29.7527737Z 85520474c47b: Pull complete 2025-07-17T08:40:29.9351312Z 4995dd45751b: Pull complete 2025-07-17T08:40:30.1108366Z 05db2817f254: Pull complete 2025-07-17T08:40:30.3093904Z 07e5f9cd2b10: Pull complete 2025-07-17T08:40:30.4628693Z 21dab22df0ec: Pull complete 2025-07-17T08:40:30.6117612Z cd89850e2c44: Pull complete 2025-07-17T08:40:30.9616008Z 7290f4ad3db1: Pull complete 2025-07-17T08:40:31.1002080Z ee114be1a250: Pull complete 2025-07-17T08:41:21.2571387Z 88d77fb451a4: Pull complete 2025-07-17T08:41:21.4558906Z eb98dfa00416: Pull complete 2025-07-17T08:41:21.6082473Z 0971ac0664a4: Pull complete 2025-07-17T08:41:27.1937834Z 9b9239f68e90: Pull complete 2025-07-17T08:41:27.3302713Z 4d3441d45503: Pull complete 2025-07-17T08:41:27.5095037Z 4f4fb700ef54: Pull complete 2025-07-17T08:41:27.6526800Z bc1ff02b1df0: Pull complete 2025-07-17T08:41:28.8922126Z e20c94532338: Pull complete 2025-07-17T08:41:28.8948619Z Digest: sha256:7dce2798440bdcaa0daa2567baeca0695de58990a0e75ad9fa0ad65b9111a6d1 2025-07-17T08:41:28.8956135Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:41:28.8964282Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:41:28.9029829Z Prepare all required actions 2025-07-17T08:41:28.9060153Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-07-17T08:41:28.9060623Z with: 2025-07-17T08:41:28.9060958Z github-token: *** 2025-07-17T08:41:28.9061123Z env: 2025-07-17T08:41:28.9061277Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:41:28.9061505Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:41:28.9061797Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:41:28.9062067Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:41:28.9062760Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:41:28.9063412Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:41:28.9063603Z AWS_REGION: us-east-1 2025-07-17T08:41:28.9063841Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:41:28.9064094Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:41:28.9067884Z AWS_SESSION_TOKEN: *** 2025-07-17T08:41:28.9068062Z ##[endgroup] 2025-07-17T08:41:28.9082479Z ##[group]Run set -eux 2025-07-17T08:41:28.9082666Z set -eux 2025-07-17T08:41:28.9082955Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-07-17T08:41:28.9095240Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:41:28.9095486Z env: 2025-07-17T08:41:28.9095637Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:41:28.9095857Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:41:28.9096144Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:41:28.9096418Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:41:28.9097105Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:41:28.9097757Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:41:28.9097951Z AWS_REGION: us-east-1 2025-07-17T08:41:28.9098170Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:41:28.9098453Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:41:28.9102256Z AWS_SESSION_TOKEN: *** 2025-07-17T08:41:28.9102500Z GITHUB_TOKEN: *** 2025-07-17T08:41:28.9102665Z ##[endgroup] 2025-07-17T08:41:28.9136976Z + python3 .github/scripts/get_workflow_job_id.py 16337959895 linux.rocm.gpu.mi300.2-8zrv9-runner-r2whm 2025-07-17T08:41:29.9867289Z Setting output job-id=46160759546 2025-07-17T08:41:29.9867789Z Setting output job-name=linux-noble-rocm-py3.12-mi300 / test (default, 2, 6, linux.rocm.gpu.mi300.2, unstable) 2025-07-17T08:41:30.0003562Z Prepare all required actions 2025-07-17T08:41:30.0003839Z Getting action download info 2025-07-17T08:41:30.3430536Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-07-17T08:41:30.7999766Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-07-17T08:41:31.2618568Z ##[group]Run ./.github/actions/download-build-artifacts 2025-07-17T08:41:31.2618817Z with: 2025-07-17T08:41:31.2618991Z name: linux-noble-rocm-py3.12-mi300 2025-07-17T08:41:31.2619201Z s3-bucket: gha-artifacts 2025-07-17T08:41:31.2619376Z env: 2025-07-17T08:41:31.2619522Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:41:31.2619739Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:41:31.2620027Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:41:31.2620295Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:41:31.2621009Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:41:31.2621680Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:41:31.2621868Z AWS_REGION: us-east-1 2025-07-17T08:41:31.2622101Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:41:31.2622670Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:41:31.2627215Z AWS_SESSION_TOKEN: *** 2025-07-17T08:41:31.2627469Z ##[endgroup] 2025-07-17T08:41:31.2657091Z ##[group]Run seemethere/download-artifact-s3@v4 2025-07-17T08:41:31.2657313Z with: 2025-07-17T08:41:31.2657486Z name: linux-noble-rocm-py3.12-mi300 2025-07-17T08:41:31.2657704Z s3-bucket: gha-artifacts 2025-07-17T08:41:31.2657895Z region: us-east-1 2025-07-17T08:41:31.2658049Z env: 2025-07-17T08:41:31.2658193Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:41:31.2658422Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:41:31.2658714Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:41:31.2658987Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:41:31.2659673Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:41:31.2660339Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:41:31.2660537Z AWS_REGION: us-east-1 2025-07-17T08:41:31.2660749Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:41:31.2661013Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:41:31.2664827Z AWS_SESSION_TOKEN: *** 2025-07-17T08:41:31.2665017Z ##[endgroup] 2025-07-17T08:41:31.6559466Z (node:5057) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-07-17T08:41:31.6559791Z 2025-07-17T08:41:31.6559957Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-07-17T08:41:31.6560309Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-07-17T08:41:31.6560677Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-07-17T08:41:31.7822012Z Found 1 objects with prefix pytorch/pytorch/16337959895/linux-noble-rocm-py3.12-mi300/ 2025-07-17T08:41:31.7822475Z Starting download (1/1): /home/runner/_work/pytorch/pytorch/artifacts.zip 2025-07-17T08:42:04.4637556Z Finished download (1/1): /home/runner/_work/pytorch/pytorch/artifacts.zip 2025-07-17T08:42:04.4642274Z Artifact download has finished successfully 2025-07-17T08:42:04.4938421Z ##[group]Run unzip -o artifacts.zip 2025-07-17T08:42:04.4938665Z unzip -o artifacts.zip 2025-07-17T08:42:04.4951055Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:42:04.4951306Z env: 2025-07-17T08:42:04.4951471Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:04.4951692Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:04.4952304Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:04.4952582Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:04.4953271Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:04.4954009Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:04.4954197Z AWS_REGION: us-east-1 2025-07-17T08:42:04.4954439Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:04.4954695Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:04.4958482Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:04.4958674Z ##[endgroup] 2025-07-17T08:42:04.5008598Z Archive: artifacts.zip 2025-07-17T08:42:04.5011739Z creating: dist/ 2025-07-17T08:42:09.7940176Z inflating: dist/torch-2.9.0a0+gita38f433-cp312-cp312-linux_x86_64.whl 2025-07-17T08:42:09.8055222Z inflating: dist/.ninja_log 2025-07-17T08:42:09.8056508Z creating: build/custom_test_artifacts/ 2025-07-17T08:42:09.8057076Z creating: build/custom_test_artifacts/custom-op-build/ 2025-07-17T08:42:09.8057933Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-07-17T08:42:09.8058765Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-07-17T08:42:09.8063646Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-07-17T08:42:09.8066823Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/ 2025-07-17T08:42:09.8067289Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeSystem.cmake 2025-07-17T08:42:09.8067741Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/ 2025-07-17T08:42:09.8068166Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/tmp/ 2025-07-17T08:42:09.8070501Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/CMakeCCompilerId.c 2025-07-17T08:42:09.8072327Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdC/a.out 2025-07-17T08:42:09.8073672Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeCCompiler.cmake 2025-07-17T08:42:09.8074334Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/ 2025-07-17T08:42:09.8075271Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/tmp/ 2025-07-17T08:42:09.8078282Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-07-17T08:42:09.8080160Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CompilerIdCXX/a.out 2025-07-17T08:42:09.8081674Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeCXXCompiler.cmake 2025-07-17T08:42:09.8083542Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_C.bin 2025-07-17T08:42:09.8085477Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_CXX.bin 2025-07-17T08:42:09.8086063Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-07-17T08:42:09.8086833Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeTmp/ 2025-07-17T08:42:09.8088307Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-07-17T08:42:09.8088858Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-07-17T08:42:09.8090389Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-07-17T08:42:09.8091501Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-07-17T08:42:09.8093097Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-07-17T08:42:09.8093963Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-07-17T08:42:09.8095268Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-07-17T08:42:09.8096710Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-07-17T08:42:09.8097977Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-07-17T08:42:09.8099272Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-07-17T08:42:09.8100356Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-07-17T08:42:09.8118540Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-07-17T08:42:09.8280310Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-07-17T08:42:09.8281236Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.d 2025-07-17T08:42:09.8281947Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-07-17T08:42:09.8283431Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-07-17T08:42:09.8284508Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-07-17T08:42:09.8285762Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-07-17T08:42:09.8287043Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-07-17T08:42:09.8288403Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-07-17T08:42:09.8289891Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-07-17T08:42:09.8291026Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-07-17T08:42:09.8292441Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-07-17T08:42:09.8293656Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-07-17T08:42:09.8311679Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-07-17T08:42:09.8380411Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-07-17T08:42:09.8381915Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.d 2025-07-17T08:42:09.8382886Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-07-17T08:42:09.8384105Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-07-17T08:42:09.8385128Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-07-17T08:42:09.8386615Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-07-17T08:42:09.8388651Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-07-17T08:42:09.8389910Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/InstallScripts.json 2025-07-17T08:42:09.8390963Z inflating: build/custom_test_artifacts/custom-op-build/hipblaslt_test_outer_vec.cc 2025-07-17T08:42:09.8392136Z inflating: build/custom_test_artifacts/custom-op-build/hipblaslt_test_vec_ext.cc 2025-07-17T08:42:09.8394348Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-07-17T08:42:09.8395936Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-07-17T08:42:09.8397269Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-07-17T08:42:09.8535383Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-07-17T08:42:09.8583681Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-07-17T08:42:09.8584051Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-07-17T08:42:09.8584678Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-07-17T08:42:09.8585666Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-07-17T08:42:09.8589997Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-07-17T08:42:09.8590444Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/ 2025-07-17T08:42:09.8591997Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeSystem.cmake 2025-07-17T08:42:09.8592681Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/ 2025-07-17T08:42:09.8593512Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/tmp/ 2025-07-17T08:42:09.8596536Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/CMakeCCompilerId.c 2025-07-17T08:42:09.8598414Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdC/a.out 2025-07-17T08:42:09.8599692Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeCCompiler.cmake 2025-07-17T08:42:09.8600534Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/ 2025-07-17T08:42:09.8601428Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/tmp/ 2025-07-17T08:42:09.8604471Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-07-17T08:42:09.8606364Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CompilerIdCXX/a.out 2025-07-17T08:42:09.8607870Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeCXXCompiler.cmake 2025-07-17T08:42:09.8609801Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_C.bin 2025-07-17T08:42:09.8611657Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_CXX.bin 2025-07-17T08:42:09.8612141Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-07-17T08:42:09.8612941Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeTmp/ 2025-07-17T08:42:09.8614418Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-07-17T08:42:09.8614929Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-07-17T08:42:09.8616483Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-07-17T08:42:09.8617607Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-07-17T08:42:09.8618832Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-07-17T08:42:09.8620075Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-07-17T08:42:09.8621181Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-07-17T08:42:09.8622730Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-07-17T08:42:09.8624013Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-07-17T08:42:09.8625283Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-07-17T08:42:09.8626530Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-07-17T08:42:09.8644901Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-07-17T08:42:09.8698918Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-07-17T08:42:09.8700438Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.d 2025-07-17T08:42:09.8701770Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-07-17T08:42:09.8702835Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-07-17T08:42:09.8704024Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-07-17T08:42:09.8705353Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-07-17T08:42:09.8707269Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-07-17T08:42:09.8708452Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/InstallScripts.json 2025-07-17T08:42:09.8709503Z inflating: build/custom_test_artifacts/jit-hook-build/hipblaslt_test_outer_vec.cc 2025-07-17T08:42:09.8710671Z inflating: build/custom_test_artifacts/jit-hook-build/hipblaslt_test_vec_ext.cc 2025-07-17T08:42:09.8712863Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-07-17T08:42:09.8714451Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-07-17T08:42:09.8715658Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-07-17T08:42:09.8749158Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-07-17T08:42:09.8749668Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-07-17T08:42:09.8750497Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-07-17T08:42:09.8751386Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-07-17T08:42:09.8756037Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-07-17T08:42:09.8756575Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/ 2025-07-17T08:42:09.8758120Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeSystem.cmake 2025-07-17T08:42:09.8758672Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/ 2025-07-17T08:42:09.8759579Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/tmp/ 2025-07-17T08:42:09.8762595Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/CMakeCCompilerId.c 2025-07-17T08:42:09.8764481Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdC/a.out 2025-07-17T08:42:09.8765807Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeCCompiler.cmake 2025-07-17T08:42:09.8766377Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/ 2025-07-17T08:42:09.8767298Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/tmp/ 2025-07-17T08:42:09.8770299Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-07-17T08:42:09.8772220Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CompilerIdCXX/a.out 2025-07-17T08:42:09.8773738Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeCXXCompiler.cmake 2025-07-17T08:42:09.8775647Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_C.bin 2025-07-17T08:42:09.8777575Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/4.0.0/CMakeDetermineCompilerABI_CXX.bin 2025-07-17T08:42:09.8778085Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-07-17T08:42:09.8779101Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeTmp/ 2025-07-17T08:42:09.8780380Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-07-17T08:42:09.8780947Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-07-17T08:42:09.8782571Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-07-17T08:42:09.8783682Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-07-17T08:42:09.8784859Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-07-17T08:42:09.8786120Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-07-17T08:42:09.8787372Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-07-17T08:42:09.8788818Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-07-17T08:42:09.8789963Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-07-17T08:42:09.8791391Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-07-17T08:42:09.8792594Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-07-17T08:42:09.8797181Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-07-17T08:42:09.8899522Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-07-17T08:42:09.8900704Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.d 2025-07-17T08:42:09.8901462Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-07-17T08:42:09.8902885Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-07-17T08:42:09.8903974Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-07-17T08:42:09.8905133Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-07-17T08:42:09.8906530Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-07-17T08:42:09.8907804Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-07-17T08:42:09.8909275Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-07-17T08:42:09.8910389Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-07-17T08:42:09.8911711Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-07-17T08:42:09.8912804Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-07-17T08:42:09.8931013Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-07-17T08:42:09.8978042Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-07-17T08:42:09.8979501Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.d 2025-07-17T08:42:09.8980447Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-07-17T08:42:09.8981788Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-07-17T08:42:09.8982666Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-07-17T08:42:09.8984219Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-07-17T08:42:09.8986137Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-07-17T08:42:09.8987318Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/InstallScripts.json 2025-07-17T08:42:09.8988339Z inflating: build/custom_test_artifacts/custom-backend-build/hipblaslt_test_outer_vec.cc 2025-07-17T08:42:09.8989544Z inflating: build/custom_test_artifacts/custom-backend-build/hipblaslt_test_vec_ext.cc 2025-07-17T08:42:09.9004380Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-07-17T08:42:09.9004799Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-07-17T08:42:09.9005180Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-07-17T08:42:09.9081829Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-07-17T08:42:09.9115847Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-07-17T08:42:09.9116152Z creating: build/lib/ 2025-07-17T08:42:09.9187660Z inflating: build/lib/libprotobuf-lite.a 2025-07-17T08:42:09.9567920Z inflating: build/lib/libprotobuf.a 2025-07-17T08:42:09.9993107Z inflating: build/lib/libprotoc.a 2025-07-17T08:42:10.0001931Z inflating: build/lib/libpthreadpool.a 2025-07-17T08:42:10.0009494Z inflating: build/lib/libcpuinfo.a 2025-07-17T08:42:10.0016324Z inflating: build/lib/libcpuinfo_internals.a 2025-07-17T08:42:10.0017705Z inflating: build/lib/libclog.a 2025-07-17T08:42:10.0034578Z inflating: build/lib/libpytorch_qnnpack.a 2025-07-17T08:42:10.0037029Z inflating: build/lib/libnnpack_reference_layers.a 2025-07-17T08:42:10.0197226Z inflating: build/lib/libmicrokernels-prod.a 2025-07-17T08:42:10.0212874Z inflating: build/lib/libnnpack.a 2025-07-17T08:42:10.0963268Z inflating: build/lib/libmicrokernels-all.a 2025-07-17T08:42:10.1022898Z inflating: build/lib/libgtest.a 2025-07-17T08:42:10.1038158Z inflating: build/lib/libgmock.a 2025-07-17T08:42:10.1039055Z inflating: build/lib/libgtest_main.a 2025-07-17T08:42:10.1040387Z inflating: build/lib/libgmock_main.a 2025-07-17T08:42:10.1118642Z inflating: build/lib/libXNNPACK.a 2025-07-17T08:42:10.1183856Z inflating: build/lib/libbenchmark.a 2025-07-17T08:42:10.1184688Z inflating: build/lib/libbenchmark_main.a 2025-07-17T08:42:10.1185989Z inflating: build/lib/libjitprofiling.a 2025-07-17T08:42:10.1242941Z inflating: build/lib/libasmjit.a 2025-07-17T08:42:10.2243987Z inflating: build/lib/libfbgemm.a 2025-07-17T08:42:10.2250803Z inflating: build/lib/libittnotify.a 2025-07-17T08:42:10.2274930Z inflating: build/lib/libtensorpipe_uv.a 2025-07-17T08:42:10.2747806Z inflating: build/lib/libtensorpipe.a 2025-07-17T08:42:10.2856058Z inflating: build/lib/libgloo.a 2025-07-17T08:42:10.2897396Z inflating: build/lib/libonnx_proto.a 2025-07-17T08:42:10.3275170Z inflating: build/lib/libgloo_hip.a 2025-07-17T08:42:10.3894033Z inflating: build/lib/libonnx.a 2025-07-17T08:42:11.2674727Z inflating: build/lib/libdnnl.a 2025-07-17T08:42:11.2691373Z inflating: build/lib/libfmt.a 2025-07-17T08:42:11.2941821Z inflating: build/lib/libkineto.a 2025-07-17T08:42:11.3034816Z inflating: build/lib/libc10.so 2025-07-17T08:42:11.3036683Z inflating: build/lib/libtorch_global_deps.so 2025-07-17T08:42:11.3038555Z inflating: build/lib/libcaffe2_nvrtc.so 2025-07-17T08:42:11.3084224Z inflating: build/lib/libc10_hip.so 2025-07-17T08:42:14.1740662Z inflating: build/lib/libtorch_cpu.so 2025-07-17T08:42:14.1744540Z inflating: build/lib/libshm.so 2025-07-17T08:42:14.9051738Z inflating: build/lib/libtorch_hip.so 2025-07-17T08:42:14.9053833Z inflating: build/lib/libtorch.so 2025-07-17T08:42:14.9069709Z inflating: build/lib/libjitbackend_test.so 2025-07-17T08:42:14.9091259Z inflating: build/lib/libbackend_with_compiler.so 2025-07-17T08:42:14.9151313Z inflating: build/lib/libtorchbind_test.so 2025-07-17T08:42:14.9174955Z inflating: build/lib/libaoti_custom_ops.so 2025-07-17T08:42:15.0901879Z inflating: build/lib/libtorch_python.so 2025-07-17T08:42:15.0931105Z inflating: build/lib/libnnapi_backend.so 2025-07-17T08:42:15.0932340Z creating: build/bin/ 2025-07-17T08:42:15.0933144Z creating: build/bin/CMakeFiles/ 2025-07-17T08:42:15.0934714Z inflating: build/bin/cmake_install.cmake 2025-07-17T08:42:15.0935956Z inflating: build/bin/CTestTestfile.cmake 2025-07-17T08:42:15.1328993Z inflating: build/bin/protoc-3.13.0.0 2025-07-17T08:42:15.1720646Z inflating: build/bin/protoc 2025-07-17T08:42:15.1768890Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-07-17T08:42:15.1818589Z inflating: build/bin/c10_DeviceGuard_test 2025-07-17T08:42:15.1868380Z inflating: build/bin/c10_Device_test 2025-07-17T08:42:15.1925357Z inflating: build/bin/c10_DispatchKeySet_test 2025-07-17T08:42:15.1973188Z inflating: build/bin/c10_StreamGuard_test 2025-07-17T08:42:15.2025007Z inflating: build/bin/c10_Scalar_test 2025-07-17T08:42:15.2074337Z inflating: build/bin/c10_SymInt_test 2025-07-17T08:42:15.2126683Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-07-17T08:42:15.2179966Z inflating: build/bin/c10_InlineStreamGuard_test 2025-07-17T08:42:15.2233769Z inflating: build/bin/c10_SizesAndStrides_test 2025-07-17T08:42:15.2300471Z inflating: build/bin/c10_cow_test 2025-07-17T08:42:15.2351303Z inflating: build/bin/c10_Bitset_test 2025-07-17T08:42:15.2399751Z inflating: build/bin/c10_ArrayRef_test 2025-07-17T08:42:15.2447610Z inflating: build/bin/c10_ConstexprCrc_test 2025-07-17T08:42:15.2495729Z inflating: build/bin/c10_DeadlockDetection_test 2025-07-17T08:42:15.2550161Z inflating: build/bin/c10_Enumerate_test 2025-07-17T08:42:15.2598573Z inflating: build/bin/c10_Half_test 2025-07-17T08:42:15.2652229Z inflating: build/bin/c10_LeftRight_test 2025-07-17T08:42:15.2703088Z inflating: build/bin/c10_IntrusiveList_test 2025-07-17T08:42:15.2756483Z inflating: build/bin/c10_Metaprogramming_test 2025-07-17T08:42:15.2807726Z inflating: build/bin/c10_NetworkFlow_test 2025-07-17T08:42:15.2855894Z inflating: build/bin/c10_Semaphore_test 2025-07-17T08:42:15.2904230Z inflating: build/bin/c10_Synchronized_test 2025-07-17T08:42:15.2957394Z inflating: build/bin/c10_ThreadLocal_test 2025-07-17T08:42:15.3006909Z inflating: build/bin/c10_TypeIndex_test 2025-07-17T08:42:15.3056337Z inflating: build/bin/c10_TypeList_test 2025-07-17T08:42:15.3103769Z inflating: build/bin/c10_TypeTraits_test 2025-07-17T08:42:15.3153615Z inflating: build/bin/c10_accumulate_test 2025-07-17T08:42:15.3207265Z inflating: build/bin/c10_bfloat16_test 2025-07-17T08:42:15.3256009Z inflating: build/bin/c10_bit_cast_test 2025-07-17T08:42:15.3310177Z inflating: build/bin/c10_complex_math_test 2025-07-17T08:42:15.3363379Z inflating: build/bin/c10_complex_test 2025-07-17T08:42:15.3411550Z inflating: build/bin/c10_error_test 2025-07-17T08:42:15.3461484Z inflating: build/bin/c10_exception_test 2025-07-17T08:42:15.3510325Z inflating: build/bin/c10_flags_test 2025-07-17T08:42:15.3559090Z inflating: build/bin/c10_generic_math_test 2025-07-17T08:42:15.3607838Z inflating: build/bin/c10_irange_test 2025-07-17T08:42:15.3756831Z inflating: build/bin/c10_intrusive_ptr_test 2025-07-17T08:42:15.3807894Z inflating: build/bin/c10_lazy_test 2025-07-17T08:42:15.3862341Z inflating: build/bin/c10_logging_test 2025-07-17T08:42:15.3933139Z inflating: build/bin/c10_optional_test 2025-07-17T08:42:15.3984179Z inflating: build/bin/c10_registry_test 2025-07-17T08:42:15.4043204Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-07-17T08:42:15.4182048Z inflating: build/bin/c10_small_vector_test 2025-07-17T08:42:15.4231906Z inflating: build/bin/c10_ssize_test 2025-07-17T08:42:15.4286306Z inflating: build/bin/c10_string_util_test 2025-07-17T08:42:15.4333996Z inflating: build/bin/c10_string_view_test 2025-07-17T08:42:15.4382624Z inflating: build/bin/c10_tempfile_test 2025-07-17T08:42:15.4436630Z inflating: build/bin/c10_typeid_test 2025-07-17T08:42:15.4478995Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-07-17T08:42:15.4527057Z inflating: build/bin/c10_hip_HIPAssertionsTest_1_var_test 2025-07-17T08:42:15.4574789Z inflating: build/bin/c10_hip_HIPAssertionsTest_catches_stream 2025-07-17T08:42:15.4624008Z inflating: build/bin/c10_hip_HIPAssertionsTest_catches_thread_and_block_and_device 2025-07-17T08:42:15.4671522Z inflating: build/bin/c10_hip_HIPAssertionsTest_from_2_processes 2025-07-17T08:42:15.4719301Z inflating: build/bin/c10_hip_HIPAssertionsTest_multiple_writes_from_blocks_and_threads 2025-07-17T08:42:15.4766902Z inflating: build/bin/c10_hip_HIPAssertionsTest_multiple_writes_from_multiple_blocks 2025-07-17T08:42:15.4814905Z inflating: build/bin/c10_hip_HIPAssertionsTest_multiple_writes_from_same_block 2025-07-17T08:42:15.4863964Z inflating: build/bin/c10_hip_HIPTest 2025-07-17T08:42:15.5861692Z inflating: build/bin/vec_test_all_types_AVX512 2025-07-17T08:42:15.6814211Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-07-17T08:42:15.7785962Z inflating: build/bin/vec_test_all_types_AVX2 2025-07-17T08:42:15.7836141Z inflating: build/bin/BackoffTest 2025-07-17T08:42:15.7887538Z inflating: build/bin/FileStoreTest 2025-07-17T08:42:15.7941915Z inflating: build/bin/TCPStoreTest 2025-07-17T08:42:15.7993966Z inflating: build/bin/HashStoreTest 2025-07-17T08:42:15.8056400Z inflating: build/bin/ProcessGroupGlooTest 2025-07-17T08:42:15.8059564Z inflating: build/bin/example_allreduce 2025-07-17T08:42:15.8063526Z inflating: build/bin/torch_shm_manager 2025-07-17T08:42:15.8114984Z inflating: build/bin/static_runtime_bench 2025-07-17T08:42:15.8165173Z inflating: build/bin/Dimname_test 2025-07-17T08:42:15.8227074Z inflating: build/bin/MaybeOwned_test 2025-07-17T08:42:15.8296124Z inflating: build/bin/Dict_test 2025-07-17T08:42:15.8352172Z inflating: build/bin/apply_utils_test 2025-07-17T08:42:15.8406708Z inflating: build/bin/NamedTensor_test 2025-07-17T08:42:15.8462800Z inflating: build/bin/atest 2025-07-17T08:42:15.8515982Z inflating: build/bin/broadcast_test 2025-07-17T08:42:15.8741822Z inflating: build/bin/static_runtime_test 2025-07-17T08:42:15.8801845Z inflating: build/bin/basic 2025-07-17T08:42:15.8850998Z inflating: build/bin/cpu_allocator_test 2025-07-17T08:42:15.8906498Z inflating: build/bin/cpu_generator_test 2025-07-17T08:42:15.8957455Z inflating: build/bin/cpu_profiling_allocator_test 2025-07-17T08:42:15.9006623Z inflating: build/bin/dlconvertor_test 2025-07-17T08:42:15.9062196Z inflating: build/bin/extension_backend_test 2025-07-17T08:42:15.9114944Z inflating: build/bin/half_test 2025-07-17T08:42:15.9163544Z inflating: build/bin/lazy_tensor_test 2025-07-17T08:42:15.9214898Z inflating: build/bin/math_kernel_test 2025-07-17T08:42:15.9266426Z inflating: build/bin/memory_format_test 2025-07-17T08:42:15.9318048Z inflating: build/bin/memory_overlapping_test 2025-07-17T08:42:15.9367278Z inflating: build/bin/operator_name_test 2025-07-17T08:42:15.9415975Z inflating: build/bin/operators_test 2025-07-17T08:42:15.9469833Z inflating: build/bin/native_test 2025-07-17T08:42:15.9555280Z inflating: build/bin/cpu_rng_test 2025-07-17T08:42:15.9605466Z inflating: build/bin/packedtensoraccessor_test 2025-07-17T08:42:15.9694411Z inflating: build/bin/ivalue_test 2025-07-17T08:42:15.9745557Z inflating: build/bin/mobile_memory_cleanup 2025-07-17T08:42:15.9800509Z inflating: build/bin/quantized_test 2025-07-17T08:42:15.9850114Z inflating: build/bin/reportMemoryUsage_test 2025-07-17T08:42:15.9897840Z inflating: build/bin/reduce_ops_test 2025-07-17T08:42:15.9961413Z inflating: build/bin/pow_test 2025-07-17T08:42:16.0015788Z inflating: build/bin/scalar_tensor_test 2025-07-17T08:42:16.0071985Z inflating: build/bin/scalar_test 2025-07-17T08:42:16.0121284Z inflating: build/bin/StorageUtils_test 2025-07-17T08:42:16.0171162Z inflating: build/bin/stride_properties_test 2025-07-17T08:42:16.0223791Z inflating: build/bin/type_ptr_test 2025-07-17T08:42:16.0276119Z inflating: build/bin/test_parallel 2025-07-17T08:42:16.0278193Z inflating: build/bin/verify_api_visibility 2025-07-17T08:42:16.0281001Z inflating: build/bin/thread_init_test 2025-07-17T08:42:16.0331195Z inflating: build/bin/undefined_tensor_test 2025-07-17T08:42:16.0405472Z inflating: build/bin/tensor_iterator_test 2025-07-17T08:42:16.0454745Z inflating: build/bin/weakref_test 2025-07-17T08:42:16.0504506Z inflating: build/bin/wrapdim_test 2025-07-17T08:42:16.0561440Z inflating: build/bin/IListRef_test 2025-07-17T08:42:16.0627630Z inflating: build/bin/legacy_vmap_test 2025-07-17T08:42:16.0677456Z inflating: build/bin/xla_tensor_test 2025-07-17T08:42:16.0734317Z inflating: build/bin/type_test 2025-07-17T08:42:16.0796920Z inflating: build/bin/KernelFunction_test 2025-07-17T08:42:16.0885218Z inflating: build/bin/kernel_function_test 2025-07-17T08:42:16.0983414Z inflating: build/bin/List_test 2025-07-17T08:42:16.1032838Z inflating: build/bin/CppSignature_test 2025-07-17T08:42:16.1090659Z inflating: build/bin/kernel_stackbased_test 2025-07-17T08:42:16.1200826Z inflating: build/bin/kernel_function_legacy_test 2025-07-17T08:42:16.1248385Z inflating: build/bin/op_allowlist_test 2025-07-17T08:42:16.1311880Z inflating: build/bin/inline_container_test 2025-07-17T08:42:16.1426600Z inflating: build/bin/kernel_lambda_legacy_test 2025-07-17T08:42:16.1477689Z inflating: build/bin/hip_apply_test 2025-07-17T08:42:16.1525143Z inflating: build/bin/hip_complex_math_test 2025-07-17T08:42:16.1572490Z inflating: build/bin/hip_complex_test 2025-07-17T08:42:16.1619762Z inflating: build/bin/hip_distributions_test 2025-07-17T08:42:16.1666893Z inflating: build/bin/hip_generator_test 2025-07-17T08:42:16.1760831Z inflating: build/bin/kernel_lambda_test 2025-07-17T08:42:16.1813667Z inflating: build/bin/backend_fallback_test 2025-07-17T08:42:16.1861317Z inflating: build/bin/hip_half_test 2025-07-17T08:42:16.1908762Z inflating: build/bin/hip_integer_divider_test 2025-07-17T08:42:16.1956078Z inflating: build/bin/hip_optional_test 2025-07-17T08:42:16.2003500Z inflating: build/bin/hip_packedtensoraccessor_test 2025-07-17T08:42:16.2051159Z inflating: build/bin/hip_vectorized_test 2025-07-17T08:42:16.2139288Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-07-17T08:42:16.2188766Z inflating: build/bin/hip_dlconvertor_test 2025-07-17T08:42:16.2463624Z inflating: build/bin/op_registration_test 2025-07-17T08:42:16.3428515Z inflating: build/bin/test_jit 2025-07-17T08:42:16.3678254Z inflating: build/bin/test_nativert 2025-07-17T08:42:16.3689987Z inflating: build/bin/tutorial_tensorexpr 2025-07-17T08:42:16.3742893Z inflating: build/bin/test_dist_autograd 2025-07-17T08:42:16.3806923Z inflating: build/bin/test_cpp_rpc 2025-07-17T08:42:16.4499135Z inflating: build/bin/test_tensorexpr 2025-07-17T08:42:16.4501701Z inflating: build/bin/parallel_benchmark 2025-07-17T08:42:16.5509904Z inflating: build/bin/test_api 2025-07-17T08:42:16.5572366Z inflating: build/bin/test_mobile_nnc 2025-07-17T08:42:16.5580448Z inflating: build/bin/aot_model_compiler_test 2025-07-17T08:42:16.5885303Z inflating: build/bin/test_lazy 2025-07-17T08:42:16.5886218Z creating: .additional_ci_files/ 2025-07-17T08:42:16.6026193Z inflating: .additional_ci_files/test-times.json 2025-07-17T08:42:16.6563325Z inflating: .additional_ci_files/test-class-times.json 2025-07-17T08:42:16.6605352Z ##[group]Run rm artifacts.zip 2025-07-17T08:42:16.6605582Z rm artifacts.zip 2025-07-17T08:42:16.6617422Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:42:16.6617676Z env: 2025-07-17T08:42:16.6617834Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:16.6618057Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:16.6618367Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:16.6618635Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:16.6619322Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:16.6619980Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:16.6620171Z AWS_REGION: us-east-1 2025-07-17T08:42:16.6620425Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:16.6620674Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:16.6624477Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:16.6624647Z ##[endgroup] 2025-07-17T08:42:16.7092850Z ##[group]Run df -H 2025-07-17T08:42:16.7093092Z df -H 2025-07-17T08:42:16.7106676Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:42:16.7107299Z env: 2025-07-17T08:42:16.7107444Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:16.7107717Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:16.7108067Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:16.7108435Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:16.7109305Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:16.7110282Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:16.7110493Z AWS_REGION: us-east-1 2025-07-17T08:42:16.7110780Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:16.7111105Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:16.7116173Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:16.7116441Z ##[endgroup] 2025-07-17T08:42:16.7162897Z Filesystem Size Used Avail Use% Mounted on 2025-07-17T08:42:16.7163221Z overlay 900G 305G 556G 36% / 2025-07-17T08:42:16.7163444Z tmpfs 68M 0 68M 0% /dev 2025-07-17T08:42:16.7163656Z nvme 15T 32G 15T 1% /run 2025-07-17T08:42:16.7163880Z /dev/nvme1n1p2 900G 305G 556G 36% /etc/hostname 2025-07-17T08:42:16.7164121Z shm 68M 8.2k 68M 1% /dev/shm 2025-07-17T08:42:16.7164383Z tmpfs 2.5T 13k 2.5T 1% /run/secrets/kubernetes.io/serviceaccount 2025-07-17T08:42:16.7164656Z tmpfs 1.3T 0 1.3T 0% /proc/acpi 2025-07-17T08:42:16.7164886Z tmpfs 1.3T 0 1.3T 0% /proc/scsi 2025-07-17T08:42:16.7165220Z tmpfs 1.3T 0 1.3T 0% /sys/firmware 2025-07-17T08:42:16.7165489Z tmpfs 1.3T 0 1.3T 0% /sys/devices/virtual/powercap 2025-07-17T08:42:16.7199975Z Prepare all required actions 2025-07-17T08:42:16.7200257Z Getting action download info 2025-07-17T08:42:16.8610103Z ##[group]Run ./.github/actions/download-td-artifacts 2025-07-17T08:42:16.8610355Z with: 2025-07-17T08:42:16.8610498Z env: 2025-07-17T08:42:16.8610649Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:16.8610870Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:16.8611158Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:16.8611431Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:16.8612120Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:16.8612769Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:16.8612966Z AWS_REGION: us-east-1 2025-07-17T08:42:16.8613240Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:16.8613495Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:16.8617267Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:16.8617450Z ##[endgroup] 2025-07-17T08:42:16.8643576Z ##[group]Run seemethere/download-artifact-s3@v4 2025-07-17T08:42:16.8643808Z with: 2025-07-17T08:42:16.8643949Z name: td_results 2025-07-17T08:42:16.8644112Z s3-bucket: gha-artifacts 2025-07-17T08:42:16.8644280Z region: us-east-1 2025-07-17T08:42:16.8644430Z env: 2025-07-17T08:42:16.8644574Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:16.8644799Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:16.8645098Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:16.8645368Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:16.8646059Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:16.8646714Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:16.8646902Z AWS_REGION: us-east-1 2025-07-17T08:42:16.8647111Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:16.8647595Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:16.8651434Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:16.8651615Z ##[endgroup] 2025-07-17T08:42:17.2642566Z (node:5088) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-07-17T08:42:17.2642897Z 2025-07-17T08:42:17.2643035Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-07-17T08:42:17.2643443Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-07-17T08:42:17.2643834Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-07-17T08:42:18.0876040Z Found 1 objects with prefix pytorch/pytorch/16337959895/td_results/ 2025-07-17T08:42:18.0876485Z Starting download (1/1): /home/runner/_work/pytorch/pytorch/td_results.json 2025-07-17T08:42:18.2363297Z Finished download (1/1): /home/runner/_work/pytorch/pytorch/td_results.json 2025-07-17T08:42:18.2368760Z Artifact download has finished successfully 2025-07-17T08:42:18.2654475Z ##[group]Run mkdir -p .additional_ci_files 2025-07-17T08:42:18.2654757Z mkdir -p .additional_ci_files 2025-07-17T08:42:18.2655048Z mv td_results.json .additional_ci_files/td_results.json || true 2025-07-17T08:42:18.2666972Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:42:18.2667226Z env: 2025-07-17T08:42:18.2667392Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:18.2667616Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:18.2667912Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:18.2668193Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:18.2669243Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:18.2669930Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:18.2670138Z AWS_REGION: us-east-1 2025-07-17T08:42:18.2670410Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:18.2670668Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:18.2674560Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:18.2674761Z ##[endgroup] 2025-07-17T08:42:18.2769528Z ##[group]Run .github/scripts/parse_ref.py 2025-07-17T08:42:18.2769787Z .github/scripts/parse_ref.py 2025-07-17T08:42:18.2782640Z shell: /usr/bin/bash -e {0} 2025-07-17T08:42:18.2782851Z env: 2025-07-17T08:42:18.2783011Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:18.2783245Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:18.2783548Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:18.2783844Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:18.2784547Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:18.2785240Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:18.2785438Z AWS_REGION: us-east-1 2025-07-17T08:42:18.2785661Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:18.2785963Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:18.2789781Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:18.2789967Z ##[endgroup] 2025-07-17T08:42:18.2945602Z Setting output branch=main 2025-07-17T08:42:18.3053690Z Prepare all required actions 2025-07-17T08:42:18.3054017Z Getting action download info 2025-07-17T08:42:18.4458268Z ##[group]Run ./.github/actions/filter-test-configs 2025-07-17T08:42:18.4458519Z with: 2025-07-17T08:42:18.4458858Z github-token: *** 2025-07-17T08:42:18.4460134Z test-matrix: {"include": [{"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}]} 2025-07-17T08:42:18.4461928Z job-name: linux-noble-rocm-py3.12-mi300 / test (default, 2, 6, linux.rocm.gpu.mi300.2, unstable) 2025-07-17T08:42:18.4462269Z env: 2025-07-17T08:42:18.4462431Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:18.4462665Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:18.4462960Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:18.4463236Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:18.4463939Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:18.4464612Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:18.4464816Z AWS_REGION: us-east-1 2025-07-17T08:42:18.4465061Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:18.4465401Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:18.4469227Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:18.4469413Z ##[endgroup] 2025-07-17T08:42:18.4501883Z ##[group]Run nick-fields/retry@v3.0.0 2025-07-17T08:42:18.4502101Z with: 2025-07-17T08:42:18.4502255Z shell: bash 2025-07-17T08:42:18.4502428Z timeout_minutes: 10 2025-07-17T08:42:18.4502596Z max_attempts: 5 2025-07-17T08:42:18.4502775Z retry_wait_seconds: 30 2025-07-17T08:42:18.4503273Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.1 2025-07-17T08:42:18.4503788Z polling_interval_seconds: 1 2025-07-17T08:42:18.4503988Z warning_on_retry: true 2025-07-17T08:42:18.4504182Z continue_on_error: false 2025-07-17T08:42:18.4504362Z env: 2025-07-17T08:42:18.4504519Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:18.4504746Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:18.4505052Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:18.4505347Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:18.4506049Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:18.4506713Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:18.4506906Z AWS_REGION: us-east-1 2025-07-17T08:42:18.4507122Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:18.4507379Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:18.4511191Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:18.4511452Z GITHUB_TOKEN: *** 2025-07-17T08:42:18.4511619Z ##[endgroup] 2025-07-17T08:42:18.5132718Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.1 2025-07-17T08:42:18.7448398Z Defaulting to user installation because normal site-packages is not writeable 2025-07-17T08:42:18.8752981Z Collecting requests==2.27.1 2025-07-17T08:42:19.6181301Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-07-17T08:42:19.6308443Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 63.1/63.1 KB 5.0 MB/s eta 0:00:00 2025-07-17T08:42:19.6956162Z Collecting pyyaml==6.0.1 2025-07-17T08:42:19.7014458Z Downloading PyYAML-6.0.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (705 kB) 2025-07-17T08:42:19.7406172Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 705.5/705.5 KB 18.8 MB/s eta 0:00:00 2025-07-17T08:42:19.8778694Z Collecting charset-normalizer~=2.0.0 2025-07-17T08:42:19.8831096Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-07-17T08:42:19.9026593Z Collecting idna<4,>=2.5 2025-07-17T08:42:19.9075910Z Downloading idna-3.10-py3-none-any.whl (70 kB) 2025-07-17T08:42:19.9105091Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 70.4/70.4 KB 45.3 MB/s eta 0:00:00 2025-07-17T08:42:19.9353727Z Collecting certifi>=2017.4.17 2025-07-17T08:42:19.9402644Z Downloading certifi-2025.7.14-py3-none-any.whl (162 kB) 2025-07-17T08:42:19.9470565Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 162.7/162.7 KB 29.4 MB/s eta 0:00:00 2025-07-17T08:42:19.9858933Z Collecting urllib3<1.27,>=1.21.1 2025-07-17T08:42:19.9909583Z Downloading urllib3-1.26.20-py2.py3-none-any.whl (144 kB) 2025-07-17T08:42:19.9975980Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 144.2/144.2 KB 27.5 MB/s eta 0:00:00 2025-07-17T08:42:20.0603784Z Installing collected packages: urllib3, pyyaml, idna, charset-normalizer, certifi, requests 2025-07-17T08:42:20.2285626Z WARNING: The script normalizer is installed in '/home/runner/.local/bin' which is not on PATH. 2025-07-17T08:42:20.2286223Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-07-17T08:42:20.2562378Z Successfully installed certifi-2025.7.14 charset-normalizer-2.0.12 idna-3.10 pyyaml-6.0.1 requests-2.27.1 urllib3-1.26.20 2025-07-17T08:42:20.5138374Z Command completed after 1 attempt(s). 2025-07-17T08:42:20.5213034Z ##[group]Run set -x 2025-07-17T08:42:20.5213280Z set -x 2025-07-17T08:42:20.5213452Z  2025-07-17T08:42:20.5213711Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-07-17T08:42:20.5214038Z # in runner workspace 2025-07-17T08:42:20.5214306Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-07-17T08:42:20.5226200Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:42:20.5226450Z env: 2025-07-17T08:42:20.5226605Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:20.5226842Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:20.5227162Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:20.5227440Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:20.5228136Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:20.5228790Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:20.5228979Z AWS_REGION: us-east-1 2025-07-17T08:42:20.5229221Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:20.5229471Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:20.5233314Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:20.5233507Z ##[endgroup] 2025-07-17T08:42:20.5265688Z + python3 /home/runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-07-17T08:42:20.5393308Z Setting output branch=main 2025-07-17T08:42:20.5443024Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-07-17T08:42:20.5443359Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-07-17T08:42:20.5443591Z echo "Job name: ${JOB_NAME}" 2025-07-17T08:42:20.5443783Z  2025-07-17T08:42:20.5444032Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-07-17T08:42:20.5444378Z # in runner workspace 2025-07-17T08:42:20.5444961Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-07-17T08:42:20.5445344Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-07-17T08:42:20.5445566Z  --job-name "${JOB_NAME}" \ 2025-07-17T08:42:20.5446872Z  --test-matrix "{"include": [{"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}]}" \ 2025-07-17T08:42:20.5448480Z  --selected-test-configs "" \ 2025-07-17T08:42:20.5448707Z  --pr-number "${PR_NUMBER}" \ 2025-07-17T08:42:20.5448903Z  --tag "${TAG}" \ 2025-07-17T08:42:20.5449111Z  --event-name "${EVENT_NAME}" \ 2025-07-17T08:42:20.5449330Z  --schedule "${SCHEDULE}" \ 2025-07-17T08:42:20.5449536Z  --branch "${HEAD_BRANCH}" 2025-07-17T08:42:20.5460805Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:42:20.5461056Z env: 2025-07-17T08:42:20.5461222Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:20.5461453Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:20.5461756Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:20.5462035Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:20.5462748Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:20.5463404Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:20.5463599Z AWS_REGION: us-east-1 2025-07-17T08:42:20.5463842Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:20.5464104Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:20.5467926Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:20.5468239Z GITHUB_TOKEN: *** 2025-07-17T08:42:20.5468549Z JOB_NAME: linux-noble-rocm-py3.12-mi300 / test (default, 2, 6, linux.rocm.gpu.mi300.2, unstable) 2025-07-17T08:42:20.5468866Z PR_NUMBER: 2025-07-17T08:42:20.5469034Z TAG: 2025-07-17T08:42:20.5469184Z EVENT_NAME: push 2025-07-17T08:42:20.5469350Z SCHEDULE: 2025-07-17T08:42:20.5469499Z HEAD_BRANCH: main 2025-07-17T08:42:20.5469662Z ##[endgroup] 2025-07-17T08:42:20.5519230Z Workflow: rocm-mi300 2025-07-17T08:42:20.5519690Z Job name: linux-noble-rocm-py3.12-mi300 / test (default, 2, 6, linux.rocm.gpu.mi300.2, unstable) 2025-07-17T08:42:20.8020941Z Setting output keep-going=True 2025-07-17T08:42:20.8021266Z Setting output ci-verbose-test-logs=False 2025-07-17T08:42:20.8021517Z Setting output ci-test-showlocals=False 2025-07-17T08:42:20.8021750Z Setting output ci-no-test-timeout=False 2025-07-17T08:42:20.8021973Z Setting output ci-no-td=False 2025-07-17T08:42:20.8022182Z Setting output ci-td-distributed=False 2025-07-17T08:42:20.8022408Z Setting output is-unstable=True 2025-07-17T08:42:20.8022615Z Setting output reenabled-issues= 2025-07-17T08:42:20.8024681Z Setting output test-matrix={"include": [{"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}]} 2025-07-17T08:42:20.8026043Z Setting output is-test-matrix-empty=False 2025-07-17T08:42:20.8138415Z ##[group]Run echo "Filtered matrix:" 2025-07-17T08:42:20.8138717Z echo "Filtered matrix:" 2025-07-17T08:42:20.8140531Z echo "{"include": [{"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.mi300.2", "unstable": "unstable"}]}" 2025-07-17T08:42:20.8142528Z  2025-07-17T08:42:20.8142712Z echo 2025-07-17T08:42:20.8142933Z echo "Is the current job unstable? True" 2025-07-17T08:42:20.8143218Z  2025-07-17T08:42:20.8143381Z echo 2025-07-17T08:42:20.8143599Z echo "Is keep-going label set? True" 2025-07-17T08:42:20.8143861Z  2025-07-17T08:42:20.8144034Z echo 2025-07-17T08:42:20.8144229Z echo "Reenabled issues? " 2025-07-17T08:42:20.8157369Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:42:20.8157718Z env: 2025-07-17T08:42:20.8157931Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:20.8158225Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:20.8158632Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:20.8158973Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:20.8159841Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:20.8160800Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:20.8161071Z AWS_REGION: us-east-1 2025-07-17T08:42:20.8161406Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:20.8161741Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:20.8166140Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:20.8166327Z ##[endgroup] 2025-07-17T08:42:20.8198987Z Filtered matrix: 2025-07-17T08:42:20.8200411Z {include: [{config: default, shard: 1, num_shards: 6, runner: linux.rocm.gpu.mi300.2, unstable: unstable}, {config: default, shard: 2, num_shards: 6, runner: linux.rocm.gpu.mi300.2, unstable: unstable}, {config: default, shard: 3, num_shards: 6, runner: linux.rocm.gpu.mi300.2, unstable: unstable}, {config: default, shard: 4, num_shards: 6, runner: linux.rocm.gpu.mi300.2, unstable: unstable}, {config: default, shard: 5, num_shards: 6, runner: linux.rocm.gpu.mi300.2, unstable: unstable}, {config: default, shard: 6, num_shards: 6, runner: linux.rocm.gpu.mi300.2, unstable: unstable}]} 2025-07-17T08:42:20.8201677Z 2025-07-17T08:42:20.8201760Z Is the current job unstable? True 2025-07-17T08:42:20.8201904Z 2025-07-17T08:42:20.8201987Z Is keep-going label set? True 2025-07-17T08:42:20.8202119Z 2025-07-17T08:42:20.8202183Z Reenabled issues? 2025-07-17T08:42:20.8245765Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-07-17T08:42:20.8246132Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-07-17T08:42:20.8257887Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T08:42:20.8258142Z env: 2025-07-17T08:42:20.8258303Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:20.8258533Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:20.8258829Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:20.8259099Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:20.8259791Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:20.8260446Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:20.8260671Z AWS_REGION: us-east-1 2025-07-17T08:42:20.8260919Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:20.8261465Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:20.8265267Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:20.8265448Z JOB_TIMEOUT: 300 2025-07-17T08:42:20.8265612Z ##[endgroup] 2025-07-17T08:42:20.8344916Z ##[group]Run set -x 2025-07-17T08:42:20.8345153Z set -x 2025-07-17T08:42:20.8345315Z  2025-07-17T08:42:20.8345493Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-07-17T08:42:20.8345756Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-07-17T08:42:20.8346006Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-07-17T08:42:20.8346248Z  TEST_COMMAND=.ci/caffe2/test.sh 2025-07-17T08:42:20.8346446Z else 2025-07-17T08:42:20.8346616Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-07-17T08:42:20.8346824Z fi 2025-07-17T08:42:20.8346976Z  2025-07-17T08:42:20.8347201Z # detached container should get cleaned up by teardown_ec2_linux 2025-07-17T08:42:20.8347562Z # TODO: Stop building test binaries as part of the build phase 2025-07-17T08:42:20.8347885Z # Used for GPU_FLAG since that doesn't play nice 2025-07-17T08:42:20.8348192Z # shellcheck disable=SC2086,SC2090 2025-07-17T08:42:20.8348415Z container_name=$(docker run \ 2025-07-17T08:42:20.8348622Z  ${GPU_FLAG:-} \ 2025-07-17T08:42:20.8348808Z  -e BUILD_ENVIRONMENT \ 2025-07-17T08:42:20.8349011Z  -e PR_NUMBER \ 2025-07-17T08:42:20.8349196Z  -e GITHUB_ACTIONS \ 2025-07-17T08:42:20.8349388Z  -e GITHUB_REPOSITORY \ 2025-07-17T08:42:20.8349582Z  -e GITHUB_WORKFLOW \ 2025-07-17T08:42:20.8349787Z  -e GITHUB_JOB \ 2025-07-17T08:42:20.8349973Z  -e GITHUB_RUN_ID \ 2025-07-17T08:42:20.8350158Z  -e GITHUB_RUN_NUMBER \ 2025-07-17T08:42:20.8350364Z  -e GITHUB_RUN_ATTEMPT \ 2025-07-17T08:42:20.8350559Z  -e JOB_ID \ 2025-07-17T08:42:20.8350725Z  -e JOB_NAME \ 2025-07-17T08:42:20.8350896Z  -e BRANCH \ 2025-07-17T08:42:20.8351062Z  -e SHA1 \ 2025-07-17T08:42:20.8351224Z  -e AWS_DEFAULT_REGION \ 2025-07-17T08:42:20.8351418Z  -e IN_WHEEL_TEST \ 2025-07-17T08:42:20.8351596Z  -e SHARD_NUMBER \ 2025-07-17T08:42:20.8351779Z  -e TEST_CONFIG \ 2025-07-17T08:42:20.8351960Z  -e NUM_TEST_SHARDS \ 2025-07-17T08:42:20.8352146Z  -e REENABLED_ISSUES \ 2025-07-17T08:42:20.8352341Z  -e CONTINUE_THROUGH_ERROR \ 2025-07-17T08:42:20.8352545Z  -e VERBOSE_TEST_LOGS \ 2025-07-17T08:42:20.8352736Z  -e TEST_SHOWLOCALS \ 2025-07-17T08:42:20.8352914Z  -e NO_TEST_TIMEOUT \ 2025-07-17T08:42:20.8353088Z  -e NO_TD \ 2025-07-17T08:42:20.8353278Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-07-17T08:42:20.8353516Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-07-17T08:42:20.8353750Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-07-17T08:42:20.8354056Z  -e TESTS_TO_INCLUDE \ 2025-07-17T08:42:20.8354252Z  -e DASHBOARD_TAG \ 2025-07-17T08:42:20.8354493Z  --env-file="${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" \ 2025-07-17T08:42:20.8354766Z  --ulimit stack=10485760:83886080 \ 2025-07-17T08:42:20.8354972Z  --ulimit core=0 \ 2025-07-17T08:42:20.8355178Z  --security-opt seccomp=unconfined \ 2025-07-17T08:42:20.8355399Z  --cap-add=SYS_PTRACE \ 2025-07-17T08:42:20.8355588Z  --shm-size="8g" \ 2025-07-17T08:42:20.8355768Z  --tty \ 2025-07-17T08:42:20.8355930Z  --detach \ 2025-07-17T08:42:20.8356102Z  --name="${container_name}" \ 2025-07-17T08:42:20.8356304Z  --user jenkins \ 2025-07-17T08:42:20.8356523Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-07-17T08:42:20.8356790Z  -w /var/lib/jenkins/workspace \ 2025-07-17T08:42:20.8356999Z  "${DOCKER_IMAGE}" 2025-07-17T08:42:20.8357457Z ) 2025-07-17T08:42:20.8357620Z # save container name for later step 2025-07-17T08:42:20.8358060Z echo "CONTAINER_NAME=${container_name}" >> "$GITHUB_ENV" 2025-07-17T08:42:20.8358514Z # jenkins user does not have write permission to mounted workspace; work-around by copying within container to jenkins home 2025-07-17T08:42:20.8359089Z docker exec -t "${container_name}" sh -c "cd .. && cp -R workspace pytorch && cd pytorch && pip install dist/*.whl && ${TEST_COMMAND}" 2025-07-17T08:42:20.8371138Z shell: /usr/bin/bash -e {0} 2025-07-17T08:42:20.8371335Z env: 2025-07-17T08:42:20.8371495Z GIT_DEFAULT_BRANCH: main 2025-07-17T08:42:20.8371721Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T08:42:20.8372018Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T08:42:20.8372304Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T08:42:20.8373013Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T08:42:20.8373699Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T08:42:20.8373884Z AWS_REGION: us-east-1 2025-07-17T08:42:20.8374123Z AWS_ACCESS_KEY_ID: *** 2025-07-17T08:42:20.8374387Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T08:42:20.8378191Z AWS_SESSION_TOKEN: *** 2025-07-17T08:42:20.8378433Z BUILD_ENVIRONMENT: linux-noble-rocm-py3.12-mi300 2025-07-17T08:42:20.8378672Z PR_NUMBER: 2025-07-17T08:42:20.8378840Z GITHUB_REPOSITORY: pytorch/pytorch 2025-07-17T08:42:20.8379053Z GITHUB_WORKFLOW: rocm-mi300 2025-07-17T08:42:20.8379229Z GITHUB_JOB: test 2025-07-17T08:42:20.8379393Z GITHUB_RUN_ID: 16337959895 2025-07-17T08:42:20.8379570Z GITHUB_RUN_NUMBER: 7135 2025-07-17T08:42:20.8379750Z GITHUB_RUN_ATTEMPT: 1 2025-07-17T08:42:20.8379918Z JOB_ID: 46160759546 2025-07-17T08:42:20.8380230Z JOB_NAME: linux-noble-rocm-py3.12-mi300 / test (default, 2, 6, linux.rocm.gpu.mi300.2, unstable) 2025-07-17T08:42:20.8380544Z BRANCH: main 2025-07-17T08:42:20.8380726Z SHA1: a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:42:20.8380943Z CONTINUE_THROUGH_ERROR: True 2025-07-17T08:42:20.8381132Z VERBOSE_TEST_LOGS: False 2025-07-17T08:42:20.8381328Z TEST_SHOWLOCALS: False 2025-07-17T08:42:20.8381508Z NO_TEST_TIMEOUT: False 2025-07-17T08:42:20.8381676Z NO_TD: False 2025-07-17T08:42:20.8381828Z TEST_CONFIG: default 2025-07-17T08:42:20.8381992Z SHARD_NUMBER: 2 2025-07-17T08:42:20.8382146Z NUM_TEST_SHARDS: 6 2025-07-17T08:42:20.8382314Z REENABLED_ISSUES: 2025-07-17T08:42:20.8382750Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:42:20.8383216Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2025-07-17T08:42:20.8383427Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 0 2025-07-17T08:42:20.8383636Z TESTS_TO_INCLUDE: 2025-07-17T08:42:20.8383794Z DASHBOARD_TAG: 2025-07-17T08:42:20.8383959Z ##[endgroup] 2025-07-17T08:42:20.8414283Z + [[ default == \m\u\l\t\i\g\p\u ]] 2025-07-17T08:42:20.8414522Z + [[ linux-noble-rocm-py3.12-mi300 == *onnx* ]] 2025-07-17T08:42:20.8414759Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-07-17T08:42:20.8423614Z +++ nproc --ignore=2 2025-07-17T08:42:20.8437662Z ++ docker run --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e MAX_JOBS=254 -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e TESTS_TO_INCLUDE -e DASHBOARD_TAG --env-file=/home/runner/_work/_temp/github_env_16337959895 --ulimit stack=10485760:83886080 --ulimit core=0 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --shm-size=8g --tty --detach --name= --user jenkins -v /home/runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-noble-rocm-n-py3-01345e7669bb7198df9fce7a02a4a12ce8c84f2d 2025-07-17T08:42:21.3602207Z + container_name=dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571 2025-07-17T08:42:21.3602703Z + echo CONTAINER_NAME=dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571 2025-07-17T08:42:21.3603395Z + docker exec -t dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571 sh -c 'cd .. && cp -R workspace pytorch && cd pytorch && pip install dist/*.whl && .ci/pytorch/test.sh' 2025-07-17T08:42:41.2523846Z Processing ./dist/torch-2.9.0a0+gita38f433-cp312-cp312-linux_x86_64.whl 2025-07-17T08:42:42.0070722Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.9.0a0+gita38f433) (3.18.0) 2025-07-17T08:42:42.0071467Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.9.0a0+gita38f433) (4.14.1) 2025-07-17T08:42:42.0079343Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.9.0a0+gita38f433) (80.9.0) 2025-07-17T08:42:42.0081232Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.9.0a0+gita38f433) (1.13.3) 2025-07-17T08:42:42.0083589Z Requirement already satisfied: networkx in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.9.0a0+gita38f433) (2.8.8) 2025-07-17T08:42:42.0085380Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.9.0a0+gita38f433) (3.1.6) 2025-07-17T08:42:42.0088879Z Requirement already satisfied: fsspec in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch==2.9.0a0+gita38f433) (2025.5.1) 2025-07-17T08:42:42.0097855Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch==2.9.0a0+gita38f433) (1.3.0) 2025-07-17T08:42:42.0176723Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch==2.9.0a0+gita38f433) (3.0.2) 2025-07-17T08:42:42.2593746Z Installing collected packages: torch 2025-07-17T08:42:52.3842925Z ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. 2025-07-17T08:42:52.3843548Z timm 1.0.14 requires torchvision, which is not installed. 2025-07-17T08:42:52.3843949Z helion 0.0.9 requires filecheck, which is not installed. 2025-07-17T08:42:52.3844296Z Successfully installed torch-2.9.0a0+gita38f433 2025-07-17T08:42:52.4557654Z + export TERM=vt100 2025-07-17T08:42:52.4557905Z + TERM=vt100 2025-07-17T08:42:52.4559946Z ++ dirname .ci/pytorch/test.sh 2025-07-17T08:42:52.4572358Z + source .ci/pytorch/common.sh 2025-07-17T08:42:52.4576254Z +++ dirname .ci/pytorch/common.sh 2025-07-17T08:42:52.4584624Z ++ source .ci/pytorch/common_utils.sh 2025-07-17T08:42:52.4586055Z +++ declare -f -t trap_add 2025-07-17T08:42:52.4589089Z ++ set -ex -o pipefail 2025-07-17T08:42:52.4590176Z ++ [[ linux-noble-rocm-py3.12-mi300 == *rocm* ]] 2025-07-17T08:42:52.4590495Z ++ unset HIP_PLATFORM 2025-07-17T08:42:52.4590681Z ++ export PYTORCH_TEST_WITH_ROCM=1 2025-07-17T08:42:52.4590880Z ++ PYTORCH_TEST_WITH_ROCM=1 2025-07-17T08:42:52.4591075Z ++ BUILD_TEST_LIBTORCH=0 2025-07-17T08:42:52.4595292Z ++ dirname .ci/pytorch/test.sh 2025-07-17T08:42:52.4605143Z + source .ci/pytorch/common-build.sh 2025-07-17T08:42:52.4606718Z ++ [[ linux-noble-rocm-py3.12-mi300 != *win-* ]] 2025-07-17T08:42:52.4614441Z ++++ dirname .ci/pytorch/common-build.sh 2025-07-17T08:42:52.4624127Z +++ cd .ci/pytorch 2025-07-17T08:42:52.4624953Z +++ pwd -P 2025-07-17T08:42:52.4627351Z ++ script_dir=/var/lib/jenkins/pytorch/.ci/pytorch 2025-07-17T08:42:52.4627640Z ++ [[ linux-noble-rocm-py3.12-mi300 == *-pch* ]] 2025-07-17T08:42:52.4627843Z ++ which sccache 2025-07-17T08:42:52.4643853Z ++ [[ -z '' ]] 2025-07-17T08:42:52.4644060Z ++ unset SCCACHE_BUCKET 2025-07-17T08:42:52.4644244Z ++ unset SCCACHE_REGION 2025-07-17T08:42:52.4644414Z ++ sccache --stop-server 2025-07-17T08:42:52.4768804Z ++ true 2025-07-17T08:42:52.4768990Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-07-17T08:42:52.4781860Z ++ trap_add sccache_epilogue EXIT 2025-07-17T08:42:52.4782084Z ++ trap_add_cmd=sccache_epilogue 2025-07-17T08:42:52.4782260Z ++ shift 2025-07-17T08:42:52.4782424Z ++ for trap_add_name in "$@" 2025-07-17T08:42:52.4789943Z ++++ trap -p EXIT 2025-07-17T08:42:52.4792618Z +++ eval 'extract_trap_cmd ' 2025-07-17T08:42:52.4792805Z ++++ extract_trap_cmd 2025-07-17T08:42:52.4792972Z ++++ printf '%s\n' '' 2025-07-17T08:42:52.4793758Z +++ printf '%s\n' sccache_epilogue 2025-07-17T08:42:52.4796184Z ++ trap -- ' 2025-07-17T08:42:52.4796343Z sccache_epilogue' EXIT 2025-07-17T08:42:52.4796508Z ++ [[ -n '' ]] 2025-07-17T08:42:52.4796687Z ++ [[ linux-noble-rocm-py3.12-mi300 == *rocm* ]] 2025-07-17T08:42:52.4796948Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-07-17T08:42:52.4797187Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-07-17T08:42:52.4797352Z ++ sccache --start-server 2025-07-17T08:42:52.4818874Z sccache: Starting the server... 2025-07-17T08:42:52.5373030Z sccache: Listening on address 127.0.0.1:4226 2025-07-17T08:42:52.5391488Z ++ sccache --zero-stats 2025-07-17T08:42:52.5419825Z Statistics zeroed. 2025-07-17T08:42:52.5423754Z ++ which ccache 2025-07-17T08:42:52.5438322Z + [[ linux-noble-rocm-py3.12-mi300 != *rocm* ]] 2025-07-17T08:42:52.5438588Z + echo 'Environment variables:' 2025-07-17T08:42:52.5438797Z Environment variables: 2025-07-17T08:42:52.5438960Z + env 2025-07-17T08:42:52.5452029Z GITHUB_WORKSPACE=/home/runner/_work/pytorch/pytorch 2025-07-17T08:42:52.5452311Z CONTINUE_THROUGH_ERROR=True 2025-07-17T08:42:52.5452539Z BUILD_ENVIRONMENT=linux-noble-rocm-py3.12-mi300 2025-07-17T08:42:52.5452797Z HOSTNAME=linux.rocm.gpu.mi300.2-8zrv9-runner-r2whm 2025-07-17T08:42:52.5453178Z GITHUB_PATH=/home/runner/_work/_temp/_runner_file_commands/add_path_0c9f809e-4f67-4469-bc26-b0cb29bf31c4 2025-07-17T08:42:52.5453502Z GITHUB_ACTION=__self 2025-07-17T08:42:52.5453673Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-07-17T08:42:52.5453868Z GITHUB_RUN_NUMBER=7135 2025-07-17T08:42:52.5454031Z TEST_CONFIG=default 2025-07-17T08:42:52.5454202Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-07-17T08:42:52.5454406Z AWS_DEFAULT_REGION=us-east-1 2025-07-17T08:42:52.5454601Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-07-17T08:42:52.5454816Z GITHUB_REF_TYPE=branch 2025-07-17T08:42:52.5457841Z *** 2025-07-17T08:42:52.5457998Z GITHUB_REPOSITORY_ID=65600975 2025-07-17T08:42:52.5458190Z GITHUB_ACTIONS=true 2025-07-17T08:42:52.5458369Z SHA1=a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:42:52.5458602Z GITHUB_SHA=a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:42:52.5458932Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/rocm-mi300.yml@refs/heads/main 2025-07-17T08:42:52.5459225Z UCC_HOME=/usr 2025-07-17T08:42:52.5459428Z VERBOSE_TEST_LOGS=False 2025-07-17T08:42:52.5459596Z GITHUB_REF=refs/heads/main 2025-07-17T08:42:52.5459765Z SHARD_NUMBER=2 2025-07-17T08:42:52.5459919Z GITHUB_REF_PROTECTED=true 2025-07-17T08:42:52.5460085Z HOME=/var/lib/jenkins 2025-07-17T08:42:52.5460276Z GITHUB_API_URL=https://api.github.com 2025-07-17T08:42:52.5460495Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-07-17T08:42:52.5460689Z LANG=C.UTF-8 2025-07-17T08:42:52.5460869Z UCX_COMMIT=cc312eaa4655c0cc5c2bcd796db938f90563bcf6 2025-07-17T08:42:52.5461087Z PYTORCH_TEST_WITH_ROCM=1 2025-07-17T08:42:52.5461674Z NUM_TEST_SHARDS=6 2025-07-17T08:42:52.5461822Z UCX_HOME=/usr 2025-07-17T08:42:52.5462331Z GITHUB_STATE=/home/runner/_work/_temp/_runner_file_commands/save_state_0c9f809e-4f67-4469-bc26-b0cb29bf31c4 2025-07-17T08:42:52.5462811Z JOB_NAME=linux-noble-rocm-py3.12-mi300 / test (default, 2, 6, linux.rocm.gpu.mi300.2, unstable) 2025-07-17T08:42:52.5463150Z MAGMA_HOME=/opt/rocm/magma 2025-07-17T08:42:52.5463467Z GITHUB_ENV=/home/runner/_work/_temp/_runner_file_commands/set_env_0c9f809e-4f67-4469-bc26-b0cb29bf31c4 2025-07-17T08:42:52.5463880Z GITHUB_EVENT_PATH=/home/runner/_work/_temp/_github_workflow/event.json 2025-07-17T08:42:52.5464136Z GITHUB_EVENT_NAME=push 2025-07-17T08:42:52.5464392Z GITHUB_ACTIONS_RUNNER_EXTRA_USER_AGENT=actions-runner-controller/0.11.0 2025-07-17T08:42:52.5464651Z DASHBOARD_TAG= 2025-07-17T08:42:52.5464799Z GITHUB_RUN_ID=16337959895 2025-07-17T08:42:52.5465138Z GITHUB_STEP_SUMMARY=/home/runner/_work/_temp/_runner_file_commands/step_summary_0c9f809e-4f67-4469-bc26-b0cb29bf31c4 2025-07-17T08:42:52.5465506Z GITHUB_ACTOR=pytorchmergebot 2025-07-17T08:42:52.5465679Z PR_NUMBER= 2025-07-17T08:42:52.5465828Z GITHUB_RUN_ATTEMPT=1 2025-07-17T08:42:52.5465998Z ANACONDA_PYTHON_VERSION=3.12 2025-07-17T08:42:52.5466214Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-07-17T08:42:52.5466429Z TERM=vt100 2025-07-17T08:42:52.5466575Z INSTALLED_VISION=yes 2025-07-17T08:42:52.5466729Z BRANCH=main 2025-07-17T08:42:52.5466882Z OPENSSL_ROOT_DIR=/opt/openssl 2025-07-17T08:42:52.5467060Z TESTS_TO_INCLUDE= 2025-07-17T08:42:52.5467326Z GITHUB_ACTION_PATH=/home/runner/_work/pytorch/pytorch/./.github/actions/setup-rocm 2025-07-17T08:42:52.5467625Z GITHUB_SERVER_URL=https://github.com 2025-07-17T08:42:52.5467828Z PYTORCH_ROCM_ARCH=gfx90a;gfx942 2025-07-17T08:42:52.5468035Z UCC_COMMIT=0c0fc21559835044ab107199e334f7157d6a0d3d 2025-07-17T08:42:52.5468247Z REENABLED_ISSUES= 2025-07-17T08:42:52.5468386Z SHLVL=1 2025-07-17T08:42:52.5468520Z MAX_JOBS=254 2025-07-17T08:42:52.5468673Z GITHUB_ACTOR_ID=97764156 2025-07-17T08:42:52.5468891Z GITHUB_WORKFLOW_SHA=a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:42:52.5469134Z GITHUB_REF_NAME=main 2025-07-17T08:42:52.5469295Z ROCM_PATH=/opt/rocm 2025-07-17T08:42:52.5469443Z GITHUB_JOB=test 2025-07-17T08:42:52.5469593Z NO_TEST_TIMEOUT=False 2025-07-17T08:42:52.5469766Z GITHUB_REPOSITORY=pytorch/pytorch 2025-07-17T08:42:52.5469949Z LC_ALL=C.UTF-8 2025-07-17T08:42:52.5470107Z GITHUB_RETENTION_DAYS=90 2025-07-17T08:42:52.5470283Z OPENSSL_DIR=/opt/openssl 2025-07-17T08:42:52.5470453Z GITHUB_ACTION_REPOSITORY= 2025-07-17T08:42:52.5471038Z PATH=/opt/cache/bin:/opt/rocm/llvm/bin:/opt/rocm/opencl/bin:/opt/rocm/hip/bin:/opt/rocm/hcc/bin:/opt/rocm/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-07-17T08:42:52.5471601Z GITHUB_BASE_REF= 2025-07-17T08:42:52.5471749Z CI=true 2025-07-17T08:42:52.5471899Z GITHUB_REPOSITORY_OWNER=pytorch 2025-07-17T08:42:52.5472080Z JOB_ID=46160759546 2025-07-17T08:42:52.5472233Z GITHUB_HEAD_REF= 2025-07-17T08:42:52.5472378Z GITHUB_ACTION_REF= 2025-07-17T08:42:52.5472522Z TEST_SHOWLOCALS=False 2025-07-17T08:42:52.5472697Z GITHUB_WORKFLOW=rocm-mi300 2025-07-17T08:42:52.5472877Z DEBIAN_FRONTEND=noninteractive 2025-07-17T08:42:52.5473218Z GITHUB_OUTPUT=/home/runner/_work/_temp/_runner_file_commands/set_output_0c9f809e-4f67-4469-bc26-b0cb29bf31c4 2025-07-17T08:42:52.5473551Z NO_TD=False 2025-07-17T08:42:52.5473700Z OLDPWD=/var/lib/jenkins 2025-07-17T08:42:52.5473864Z _=/usr/bin/env 2025-07-17T08:42:52.5474205Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-07-17T08:42:52.5576066Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch 2025-07-17T08:42:52.5576519Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/bin 2025-07-17T08:42:52.5576875Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/lib 2025-07-17T08:42:52.5577225Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/test 2025-07-17T08:42:52.5577768Z + BUILD_DIR=build 2025-07-17T08:42:52.5577936Z + BUILD_RENAMED_DIR=build_renamed 2025-07-17T08:42:52.5578265Z + BUILD_BIN_DIR=build/bin 2025-07-17T08:42:52.5578434Z + SHARD_NUMBER=2 2025-07-17T08:42:52.5578585Z + NUM_TEST_SHARDS=6 2025-07-17T08:42:52.5578754Z + export TORCH_SERIALIZATION_DEBUG=1 2025-07-17T08:42:52.5578955Z + TORCH_SERIALIZATION_DEBUG=1 2025-07-17T08:42:52.5579134Z + export VALGRIND=ON 2025-07-17T08:42:52.5579296Z + VALGRIND=ON 2025-07-17T08:42:52.5579476Z + [[ linux-noble-rocm-py3.12-mi300 == *clang9* ]] 2025-07-17T08:42:52.5579710Z + [[ linux-noble-rocm-py3.12-mi300 == *xpu* ]] 2025-07-17T08:42:52.5579930Z + [[ linux-noble-rocm-py3.12-mi300 == *s390x* ]] 2025-07-17T08:42:52.5580133Z + [[ 0 == \1 ]] 2025-07-17T08:42:52.5580276Z + [[ True == \1 ]] 2025-07-17T08:42:52.5580446Z + [[ linux-noble-rocm-py3.12-mi300 != *bazel* ]] 2025-07-17T08:42:52.5581860Z ++ realpath build/custom_test_artifacts 2025-07-17T08:42:52.5594604Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/pytorch/build/custom_test_artifacts 2025-07-17T08:42:52.5594916Z + [[ -n '' ]] 2025-07-17T08:42:52.5595084Z + echo 'Environment variables' 2025-07-17T08:42:52.5595273Z Environment variables 2025-07-17T08:42:52.5595430Z + env 2025-07-17T08:42:52.5604079Z GITHUB_WORKSPACE=/home/runner/_work/pytorch/pytorch 2025-07-17T08:42:52.5604392Z CONTINUE_THROUGH_ERROR=True 2025-07-17T08:42:52.5604610Z BUILD_ENVIRONMENT=linux-noble-rocm-py3.12-mi300 2025-07-17T08:42:52.5604874Z HOSTNAME=linux.rocm.gpu.mi300.2-8zrv9-runner-r2whm 2025-07-17T08:42:52.5605265Z GITHUB_PATH=/home/runner/_work/_temp/_runner_file_commands/add_path_0c9f809e-4f67-4469-bc26-b0cb29bf31c4 2025-07-17T08:42:52.5605591Z GITHUB_ACTION=__self 2025-07-17T08:42:52.5605762Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-07-17T08:42:52.5605960Z GITHUB_RUN_NUMBER=7135 2025-07-17T08:42:52.5606125Z TEST_CONFIG=default 2025-07-17T08:42:52.5606294Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-07-17T08:42:52.5606499Z AWS_DEFAULT_REGION=us-east-1 2025-07-17T08:42:52.5606708Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-07-17T08:42:52.5606909Z GITHUB_REF_TYPE=branch 2025-07-17T08:42:52.5607124Z *** 2025-07-17T08:42:52.5607281Z GITHUB_REPOSITORY_ID=65600975 2025-07-17T08:42:52.5607469Z GITHUB_ACTIONS=true 2025-07-17T08:42:52.5607652Z SHA1=a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:42:52.5607874Z GITHUB_SHA=a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:42:52.5608198Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/rocm-mi300.yml@refs/heads/main 2025-07-17T08:42:52.5608490Z UCC_HOME=/usr 2025-07-17T08:42:52.5608641Z TORCH_SERIALIZATION_DEBUG=1 2025-07-17T08:42:52.5608816Z VERBOSE_TEST_LOGS=False 2025-07-17T08:42:52.5608987Z GITHUB_REF=refs/heads/main 2025-07-17T08:42:52.5609162Z SHARD_NUMBER=2 2025-07-17T08:42:52.5609324Z GITHUB_REF_PROTECTED=true 2025-07-17T08:42:52.5609501Z HOME=/var/lib/jenkins 2025-07-17T08:42:52.5609686Z GITHUB_API_URL=https://api.github.com 2025-07-17T08:42:52.5609901Z PYTORCH_TEST_RERUN_DISABLED_TESTS=0 2025-07-17T08:42:52.5610088Z LANG=C.UTF-8 2025-07-17T08:42:52.5610268Z UCX_COMMIT=cc312eaa4655c0cc5c2bcd796db938f90563bcf6 2025-07-17T08:42:52.5610496Z PYTORCH_TEST_WITH_ROCM=1 2025-07-17T08:42:52.5610668Z NUM_TEST_SHARDS=6 2025-07-17T08:42:52.5610820Z UCX_HOME=/usr 2025-07-17T08:42:52.5611128Z GITHUB_STATE=/home/runner/_work/_temp/_runner_file_commands/save_state_0c9f809e-4f67-4469-bc26-b0cb29bf31c4 2025-07-17T08:42:52.5611598Z JOB_NAME=linux-noble-rocm-py3.12-mi300 / test (default, 2, 6, linux.rocm.gpu.mi300.2, unstable) 2025-07-17T08:42:52.5611909Z MAGMA_HOME=/opt/rocm/magma 2025-07-17T08:42:52.5612215Z GITHUB_ENV=/home/runner/_work/_temp/_runner_file_commands/set_env_0c9f809e-4f67-4469-bc26-b0cb29bf31c4 2025-07-17T08:42:52.5612604Z GITHUB_EVENT_PATH=/home/runner/_work/_temp/_github_workflow/event.json 2025-07-17T08:42:52.5612865Z GITHUB_EVENT_NAME=push 2025-07-17T08:42:52.5613116Z GITHUB_ACTIONS_RUNNER_EXTRA_USER_AGENT=actions-runner-controller/0.11.0 2025-07-17T08:42:52.5613370Z DASHBOARD_TAG= 2025-07-17T08:42:52.5613764Z GITHUB_RUN_ID=16337959895 2025-07-17T08:42:52.5614229Z GITHUB_STEP_SUMMARY=/home/runner/_work/_temp/_runner_file_commands/step_summary_0c9f809e-4f67-4469-bc26-b0cb29bf31c4 2025-07-17T08:42:52.5614596Z GITHUB_ACTOR=pytorchmergebot 2025-07-17T08:42:52.5614771Z PR_NUMBER= 2025-07-17T08:42:52.5614915Z GITHUB_RUN_ATTEMPT=1 2025-07-17T08:42:52.5615105Z VALGRIND=ON 2025-07-17T08:42:52.5615256Z ANACONDA_PYTHON_VERSION=3.12 2025-07-17T08:42:52.5615469Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-07-17T08:42:52.5615683Z TERM=vt100 2025-07-17T08:42:52.5615824Z INSTALLED_VISION=yes 2025-07-17T08:42:52.5615971Z BRANCH=main 2025-07-17T08:42:52.5616114Z OPENSSL_ROOT_DIR=/opt/openssl 2025-07-17T08:42:52.5616288Z TESTS_TO_INCLUDE= 2025-07-17T08:42:52.5616541Z GITHUB_ACTION_PATH=/home/runner/_work/pytorch/pytorch/./.github/actions/setup-rocm 2025-07-17T08:42:52.5616846Z GITHUB_SERVER_URL=https://github.com 2025-07-17T08:42:52.5617046Z PYTORCH_ROCM_ARCH=gfx90a;gfx942 2025-07-17T08:42:52.5617258Z UCC_COMMIT=0c0fc21559835044ab107199e334f7157d6a0d3d 2025-07-17T08:42:52.5617466Z REENABLED_ISSUES= 2025-07-17T08:42:52.5617609Z SHLVL=1 2025-07-17T08:42:52.5617748Z MAX_JOBS=254 2025-07-17T08:42:52.5617899Z GITHUB_ACTOR_ID=97764156 2025-07-17T08:42:52.5618116Z GITHUB_WORKFLOW_SHA=a38f433be2e94a64b095a44ba39879d02d0c2316 2025-07-17T08:42:52.5618345Z GITHUB_REF_NAME=main 2025-07-17T08:42:52.5618501Z ROCM_PATH=/opt/rocm 2025-07-17T08:42:52.5618654Z GITHUB_JOB=test 2025-07-17T08:42:52.5618802Z NO_TEST_TIMEOUT=False 2025-07-17T08:42:52.5618976Z GITHUB_REPOSITORY=pytorch/pytorch 2025-07-17T08:42:52.5619162Z LC_ALL=C.UTF-8 2025-07-17T08:42:52.5619314Z GITHUB_RETENTION_DAYS=90 2025-07-17T08:42:52.5619482Z OPENSSL_DIR=/opt/openssl 2025-07-17T08:42:52.5619649Z GITHUB_ACTION_REPOSITORY= 2025-07-17T08:42:52.5620222Z PATH=/opt/cache/bin:/opt/rocm/llvm/bin:/opt/rocm/opencl/bin:/opt/rocm/hip/bin:/opt/rocm/hcc/bin:/opt/rocm/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-07-17T08:42:52.5620786Z GITHUB_BASE_REF= 2025-07-17T08:42:52.5620925Z CI=true 2025-07-17T08:42:52.5621069Z GITHUB_REPOSITORY_OWNER=pytorch 2025-07-17T08:42:52.5621247Z JOB_ID=46160759546 2025-07-17T08:42:52.5621392Z GITHUB_HEAD_REF= 2025-07-17T08:42:52.5621536Z GITHUB_ACTION_REF= 2025-07-17T08:42:52.5621683Z TEST_SHOWLOCALS=False 2025-07-17T08:42:52.5621849Z GITHUB_WORKFLOW=rocm-mi300 2025-07-17T08:42:52.5622030Z DEBIAN_FRONTEND=noninteractive 2025-07-17T08:42:52.5622363Z GITHUB_OUTPUT=/home/runner/_work/_temp/_runner_file_commands/set_output_0c9f809e-4f67-4469-bc26-b0cb29bf31c4 2025-07-17T08:42:52.5622685Z NO_TD=False 2025-07-17T08:42:52.5622821Z OLDPWD=/var/lib/jenkins 2025-07-17T08:42:52.5622982Z _=/usr/bin/env 2025-07-17T08:42:52.5623134Z + echo 'Testing pytorch' 2025-07-17T08:42:52.5623297Z Testing pytorch 2025-07-17T08:42:52.5623452Z + export LANG=C.UTF-8 2025-07-17T08:42:52.5623603Z + LANG=C.UTF-8 2025-07-17T08:42:52.5623745Z + PR_NUMBER= 2025-07-17T08:42:52.5623894Z + [[ default == \d\e\f\a\u\l\t ]] 2025-07-17T08:42:52.5624083Z + export CUDA_VISIBLE_DEVICES=0 2025-07-17T08:42:52.5624255Z + CUDA_VISIBLE_DEVICES=0 2025-07-17T08:42:52.5624422Z + export HIP_VISIBLE_DEVICES=0 2025-07-17T08:42:52.5624601Z + HIP_VISIBLE_DEVICES=0 2025-07-17T08:42:52.5624775Z + [[ default == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-07-17T08:42:52.5624968Z + [[ default == \s\l\o\w ]] 2025-07-17T08:42:52.5625184Z + [[ linux-noble-rocm-py3.12-mi300 == *slow-gradcheck* ]] 2025-07-17T08:42:52.5625431Z + [[ linux-noble-rocm-py3.12-mi300 == *cuda* ]] 2025-07-17T08:42:52.5625651Z + [[ linux-noble-rocm-py3.12-mi300 == *rocm* ]] 2025-07-17T08:42:52.5625866Z + export PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-07-17T08:42:52.5636333Z + PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-07-17T08:42:52.5636603Z + [[ default == *crossref* ]] 2025-07-17T08:42:52.5636812Z + [[ linux-noble-rocm-py3.12-mi300 == *rocm* ]] 2025-07-17T08:42:52.5637034Z + export VALGRIND=OFF 2025-07-17T08:42:52.5637202Z + VALGRIND=OFF 2025-07-17T08:42:52.5637345Z + rocminfo 2025-07-17T08:42:52.5789778Z ROCk module version 6.12.12 is loaded 2025-07-17T08:42:52.7834122Z ===================== 2025-07-17T08:42:52.7835045Z HSA System Attributes 2025-07-17T08:42:52.7835226Z ===================== 2025-07-17T08:42:52.7835402Z Runtime Version: 1.15 2025-07-17T08:42:52.7835580Z Runtime Ext Version: 1.7 2025-07-17T08:42:52.7835770Z System Timestamp Freq.: 1000.000000MHz 2025-07-17T08:42:52.7836085Z Sig. Max Wait Duration: 18446744073709551615 (0xFFFFFFFFFFFFFFFF) (timestamp count) 2025-07-17T08:42:52.7836426Z Machine Model: LARGE 2025-07-17T08:42:52.7836686Z System Endianness: LITTLE 2025-07-17T08:42:52.7836911Z Mwaitx: DISABLED 2025-07-17T08:42:52.7837098Z XNACK enabled: NO 2025-07-17T08:42:52.7837280Z DMAbuf Support: YES 2025-07-17T08:42:52.7837468Z VMM Support: YES 2025-07-17T08:42:52.7837574Z 2025-07-17T08:42:52.7837644Z ========== 2025-07-17T08:42:52.7837827Z HSA Agents 2025-07-17T08:42:52.7837996Z ========== 2025-07-17T08:42:52.7838189Z ******* 2025-07-17T08:42:52.7838367Z Agent 1 2025-07-17T08:42:52.7838524Z ******* 2025-07-17T08:42:52.7838723Z Name: AMD EPYC 9534 64-Core Processor 2025-07-17T08:42:52.7838969Z Uuid: CPU-XX 2025-07-17T08:42:52.7839237Z Marketing Name: AMD EPYC 9534 64-Core Processor 2025-07-17T08:42:52.7839490Z Vendor Name: CPU 2025-07-17T08:42:52.7839724Z Feature: None specified 2025-07-17T08:42:52.7839964Z Profile: FULL_PROFILE 2025-07-17T08:42:52.7840252Z Float Round Mode: NEAR 2025-07-17T08:42:52.7840497Z Max Queue Number: 0(0x0) 2025-07-17T08:42:52.7840747Z Queue Min Size: 0(0x0) 2025-07-17T08:42:52.7840987Z Queue Max Size: 0(0x0) 2025-07-17T08:42:52.7841223Z Queue Type: MULTI 2025-07-17T08:42:52.7841475Z Node: 0 2025-07-17T08:42:52.7841717Z Device Type: CPU 2025-07-17T08:42:52.7841936Z Cache Info: 2025-07-17T08:42:52.7842134Z L1: 32768(0x8000) KB 2025-07-17T08:42:52.7842362Z Chip ID: 0(0x0) 2025-07-17T08:42:52.7842600Z ASIC Revision: 0(0x0) 2025-07-17T08:42:52.7842849Z Cacheline Size: 64(0x40) 2025-07-17T08:42:52.7843097Z Max Clock Freq. (MHz): 2450 2025-07-17T08:42:52.7843328Z BDFID: 0 2025-07-17T08:42:52.7843565Z Internal Node ID: 0 2025-07-17T08:42:52.7843807Z Compute Unit: 128 2025-07-17T08:42:52.7844063Z SIMDs per CU: 0 2025-07-17T08:42:52.7844299Z Shader Engines: 0 2025-07-17T08:42:52.7844556Z Shader Arrs. per Eng.: 0 2025-07-17T08:42:52.7844803Z WatchPts on Addr. Ranges:1 2025-07-17T08:42:52.7845029Z Memory Properties: 2025-07-17T08:42:52.7845208Z Features: None 2025-07-17T08:42:52.7845386Z Pool Info: 2025-07-17T08:42:52.7845572Z Pool 1 2025-07-17T08:42:52.7845785Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-07-17T08:42:52.7846035Z Size: 1188703740(0x46da2dfc) KB 2025-07-17T08:42:52.7846486Z Allocatable: TRUE 2025-07-17T08:42:52.7846845Z Alloc Granule: 4KB 2025-07-17T08:42:52.7847115Z Alloc Recommended Granule:4KB 2025-07-17T08:42:52.7847393Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7847644Z Accessible by all: TRUE 2025-07-17T08:42:52.7847883Z Pool 2 2025-07-17T08:42:52.7848133Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-07-17T08:42:52.7848368Z Size: 1188703740(0x46da2dfc) KB 2025-07-17T08:42:52.7848597Z Allocatable: TRUE 2025-07-17T08:42:52.7848843Z Alloc Granule: 4KB 2025-07-17T08:42:52.7849196Z Alloc Recommended Granule:4KB 2025-07-17T08:42:52.7849458Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7849712Z Accessible by all: TRUE 2025-07-17T08:42:52.7849938Z Pool 3 2025-07-17T08:42:52.7850137Z Segment: GLOBAL; FLAGS: KERNARG, FINE GRAINED 2025-07-17T08:42:52.7850370Z Size: 1188703740(0x46da2dfc) KB 2025-07-17T08:42:52.7850606Z Allocatable: TRUE 2025-07-17T08:42:52.7850853Z Alloc Granule: 4KB 2025-07-17T08:42:52.7851107Z Alloc Recommended Granule:4KB 2025-07-17T08:42:52.7851363Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7851605Z Accessible by all: TRUE 2025-07-17T08:42:52.7851822Z Pool 4 2025-07-17T08:42:52.7852021Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-07-17T08:42:52.7852253Z Size: 1188703740(0x46da2dfc) KB 2025-07-17T08:42:52.7852488Z Allocatable: TRUE 2025-07-17T08:42:52.7852729Z Alloc Granule: 4KB 2025-07-17T08:42:52.7852982Z Alloc Recommended Granule:4KB 2025-07-17T08:42:52.7853238Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7853483Z Accessible by all: TRUE 2025-07-17T08:42:52.7853700Z ISA Info: 2025-07-17T08:42:52.7853865Z ******* 2025-07-17T08:42:52.7854030Z Agent 2 2025-07-17T08:42:52.7854181Z ******* 2025-07-17T08:42:52.7854370Z Name: AMD EPYC 9534 64-Core Processor 2025-07-17T08:42:52.7854598Z Uuid: CPU-XX 2025-07-17T08:42:52.7854844Z Marketing Name: AMD EPYC 9534 64-Core Processor 2025-07-17T08:42:52.7855093Z Vendor Name: CPU 2025-07-17T08:42:52.7855334Z Feature: None specified 2025-07-17T08:42:52.7855572Z Profile: FULL_PROFILE 2025-07-17T08:42:52.7855812Z Float Round Mode: NEAR 2025-07-17T08:42:52.7856052Z Max Queue Number: 0(0x0) 2025-07-17T08:42:52.7856288Z Queue Min Size: 0(0x0) 2025-07-17T08:42:52.7856516Z Queue Max Size: 0(0x0) 2025-07-17T08:42:52.7856750Z Queue Type: MULTI 2025-07-17T08:42:52.7856970Z Node: 1 2025-07-17T08:42:52.7857192Z Device Type: CPU 2025-07-17T08:42:52.7857534Z Cache Info: 2025-07-17T08:42:52.7857712Z L1: 32768(0x8000) KB 2025-07-17T08:42:52.7858034Z Chip ID: 0(0x0) 2025-07-17T08:42:52.7858265Z ASIC Revision: 0(0x0) 2025-07-17T08:42:52.7858507Z Cacheline Size: 64(0x40) 2025-07-17T08:42:52.7858747Z Max Clock Freq. (MHz): 2450 2025-07-17T08:42:52.7858975Z BDFID: 0 2025-07-17T08:42:52.7859201Z Internal Node ID: 1 2025-07-17T08:42:52.7859434Z Compute Unit: 128 2025-07-17T08:42:52.7859668Z SIMDs per CU: 0 2025-07-17T08:42:52.7859908Z Shader Engines: 0 2025-07-17T08:42:52.7860151Z Shader Arrs. per Eng.: 0 2025-07-17T08:42:52.7860409Z WatchPts on Addr. Ranges:1 2025-07-17T08:42:52.7860634Z Memory Properties: 2025-07-17T08:42:52.7860812Z Features: None 2025-07-17T08:42:52.7860987Z Pool Info: 2025-07-17T08:42:52.7861156Z Pool 1 2025-07-17T08:42:52.7861366Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-07-17T08:42:52.7861609Z Size: 1188946684(0x46dde2fc) KB 2025-07-17T08:42:52.7861842Z Allocatable: TRUE 2025-07-17T08:42:52.7862090Z Alloc Granule: 4KB 2025-07-17T08:42:52.7862349Z Alloc Recommended Granule:4KB 2025-07-17T08:42:52.7862605Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7862862Z Accessible by all: TRUE 2025-07-17T08:42:52.7863084Z Pool 2 2025-07-17T08:42:52.7863288Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-07-17T08:42:52.7863517Z Size: 1188946684(0x46dde2fc) KB 2025-07-17T08:42:52.7863788Z Allocatable: TRUE 2025-07-17T08:42:52.7864039Z Alloc Granule: 4KB 2025-07-17T08:42:52.7864290Z Alloc Recommended Granule:4KB 2025-07-17T08:42:52.7864546Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7864796Z Accessible by all: TRUE 2025-07-17T08:42:52.7865009Z Pool 3 2025-07-17T08:42:52.7865210Z Segment: GLOBAL; FLAGS: KERNARG, FINE GRAINED 2025-07-17T08:42:52.7865446Z Size: 1188946684(0x46dde2fc) KB 2025-07-17T08:42:52.7865683Z Allocatable: TRUE 2025-07-17T08:42:52.7865930Z Alloc Granule: 4KB 2025-07-17T08:42:52.7866186Z Alloc Recommended Granule:4KB 2025-07-17T08:42:52.7866439Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7866687Z Accessible by all: TRUE 2025-07-17T08:42:52.7866894Z Pool 4 2025-07-17T08:42:52.7867096Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-07-17T08:42:52.7867321Z Size: 1188946684(0x46dde2fc) KB 2025-07-17T08:42:52.7867555Z Allocatable: TRUE 2025-07-17T08:42:52.7867802Z Alloc Granule: 4KB 2025-07-17T08:42:52.7868057Z Alloc Recommended Granule:4KB 2025-07-17T08:42:52.7868442Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7868864Z Accessible by all: TRUE 2025-07-17T08:42:52.7869086Z ISA Info: 2025-07-17T08:42:52.7869252Z ******* 2025-07-17T08:42:52.7869412Z Agent 3 2025-07-17T08:42:52.7869564Z ******* 2025-07-17T08:42:52.7869733Z Name: gfx942 2025-07-17T08:42:52.7869956Z Uuid: GPU-6d96a91a4aa636e8 2025-07-17T08:42:52.7870197Z Marketing Name: AMD Instinct MI300X 2025-07-17T08:42:52.7870443Z Vendor Name: AMD 2025-07-17T08:42:52.7870676Z Feature: KERNEL_DISPATCH 2025-07-17T08:42:52.7870911Z Profile: BASE_PROFILE 2025-07-17T08:42:52.7871154Z Float Round Mode: NEAR 2025-07-17T08:42:52.7871404Z Max Queue Number: 128(0x80) 2025-07-17T08:42:52.7871647Z Queue Min Size: 64(0x40) 2025-07-17T08:42:52.7871880Z Queue Max Size: 131072(0x20000) 2025-07-17T08:42:52.7872113Z Queue Type: MULTI 2025-07-17T08:42:52.7872338Z Node: 2 2025-07-17T08:42:52.7872559Z Device Type: GPU 2025-07-17T08:42:52.7872778Z Cache Info: 2025-07-17T08:42:52.7872952Z L1: 32(0x20) KB 2025-07-17T08:42:52.7873159Z L2: 4096(0x1000) KB 2025-07-17T08:42:52.7873359Z L3: 262144(0x40000) KB 2025-07-17T08:42:52.7873571Z Chip ID: 29857(0x74a1) 2025-07-17T08:42:52.7873814Z ASIC Revision: 1(0x1) 2025-07-17T08:42:52.7874286Z Cacheline Size: 128(0x80) 2025-07-17T08:42:52.7874531Z Max Clock Freq. (MHz): 2100 2025-07-17T08:42:52.7874761Z BDFID: 42752 2025-07-17T08:42:52.7874990Z Internal Node ID: 2 2025-07-17T08:42:52.7875232Z Compute Unit: 304 2025-07-17T08:42:52.7875472Z SIMDs per CU: 4 2025-07-17T08:42:52.7875705Z Shader Engines: 32 2025-07-17T08:42:52.7875948Z Shader Arrs. per Eng.: 1 2025-07-17T08:42:52.7876203Z WatchPts on Addr. Ranges:4 2025-07-17T08:42:52.7876458Z Coherent Host Access: FALSE 2025-07-17T08:42:52.7876691Z Memory Properties: 2025-07-17T08:42:52.7876879Z Features: KERNEL_DISPATCH 2025-07-17T08:42:52.7877115Z Fast F16 Operation: TRUE 2025-07-17T08:42:52.7877369Z Wavefront Size: 64(0x40) 2025-07-17T08:42:52.7877617Z Workgroup Max Size: 1024(0x400) 2025-07-17T08:42:52.7877861Z Workgroup Max Size per Dimension: 2025-07-17T08:42:52.7878074Z x 1024(0x400) 2025-07-17T08:42:52.7878282Z y 1024(0x400) 2025-07-17T08:42:52.7878480Z z 1024(0x400) 2025-07-17T08:42:52.7878709Z Max Waves Per CU: 32(0x20) 2025-07-17T08:42:52.7878959Z Max Work-item Per CU: 2048(0x800) 2025-07-17T08:42:52.7879204Z Grid Max Size: 4294967295(0xffffffff) 2025-07-17T08:42:52.7879596Z Grid Max Size per Dimension: 2025-07-17T08:42:52.7879778Z x 4294967295(0xffffffff) 2025-07-17T08:42:52.7880117Z y 4294967295(0xffffffff) 2025-07-17T08:42:52.7880327Z z 4294967295(0xffffffff) 2025-07-17T08:42:52.7880559Z Max fbarriers/Workgrp: 32 2025-07-17T08:42:52.7880867Z Packet Processor uCode:: 177 2025-07-17T08:42:52.7881122Z SDMA engine uCode:: 24 2025-07-17T08:42:52.7881366Z IOMMU Support:: None 2025-07-17T08:42:52.7881587Z Pool Info: 2025-07-17T08:42:52.7881755Z Pool 1 2025-07-17T08:42:52.7881969Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-07-17T08:42:52.7882212Z Size: 201310208(0xbffc000) KB 2025-07-17T08:42:52.7882448Z Allocatable: TRUE 2025-07-17T08:42:52.7882706Z Alloc Granule: 4KB 2025-07-17T08:42:52.7882972Z Alloc Recommended Granule:2048KB 2025-07-17T08:42:52.7883231Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7883482Z Accessible by all: FALSE 2025-07-17T08:42:52.7883702Z Pool 2 2025-07-17T08:42:52.7883921Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-07-17T08:42:52.7884175Z Size: 201310208(0xbffc000) KB 2025-07-17T08:42:52.7884416Z Allocatable: TRUE 2025-07-17T08:42:52.7884660Z Alloc Granule: 4KB 2025-07-17T08:42:52.7884915Z Alloc Recommended Granule:2048KB 2025-07-17T08:42:52.7885178Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7885426Z Accessible by all: FALSE 2025-07-17T08:42:52.7885658Z Pool 3 2025-07-17T08:42:52.7885859Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-07-17T08:42:52.7886089Z Size: 201310208(0xbffc000) KB 2025-07-17T08:42:52.7886320Z Allocatable: TRUE 2025-07-17T08:42:52.7886562Z Alloc Granule: 4KB 2025-07-17T08:42:52.7886828Z Alloc Recommended Granule:2048KB 2025-07-17T08:42:52.7887087Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7887337Z Accessible by all: FALSE 2025-07-17T08:42:52.7887555Z Pool 4 2025-07-17T08:42:52.7887744Z Segment: GROUP 2025-07-17T08:42:52.7887971Z Size: 64(0x40) KB 2025-07-17T08:42:52.7888197Z Allocatable: FALSE 2025-07-17T08:42:52.7888442Z Alloc Granule: 0KB 2025-07-17T08:42:52.7888691Z Alloc Recommended Granule:0KB 2025-07-17T08:42:52.7888945Z Alloc Alignment: 0KB 2025-07-17T08:42:52.7889198Z Accessible by all: FALSE 2025-07-17T08:42:52.7889421Z ISA Info: 2025-07-17T08:42:52.7889590Z ISA 1 2025-07-17T08:42:52.7889795Z Name: amdgcn-amd-amdhsa--gfx942:sramecc+:xnack- 2025-07-17T08:42:52.7890064Z Machine Models: HSA_MACHINE_MODEL_LARGE 2025-07-17T08:42:52.7890321Z Profiles: HSA_PROFILE_BASE 2025-07-17T08:42:52.7890705Z Default Rounding Mode: NEAR 2025-07-17T08:42:52.7891063Z Default Rounding Mode: NEAR 2025-07-17T08:42:52.7891311Z Fast f16: TRUE 2025-07-17T08:42:52.7891549Z Workgroup Max Size: 1024(0x400) 2025-07-17T08:42:52.7891781Z Workgroup Max Size per Dimension: 2025-07-17T08:42:52.7891991Z x 1024(0x400) 2025-07-17T08:42:52.7892199Z y 1024(0x400) 2025-07-17T08:42:52.7892395Z z 1024(0x400) 2025-07-17T08:42:52.7892612Z Grid Max Size: 4294967295(0xffffffff) 2025-07-17T08:42:52.7892837Z Grid Max Size per Dimension: 2025-07-17T08:42:52.7893025Z x 4294967295(0xffffffff) 2025-07-17T08:42:52.7893233Z y 4294967295(0xffffffff) 2025-07-17T08:42:52.7893443Z z 4294967295(0xffffffff) 2025-07-17T08:42:52.7893681Z FBarrier Max Size: 32 2025-07-17T08:42:52.7893900Z ISA 2 2025-07-17T08:42:52.7894125Z Name: amdgcn-amd-amdhsa--gfx9-4-generic:sramecc+:xnack- 2025-07-17T08:42:52.7894401Z Machine Models: HSA_MACHINE_MODEL_LARGE 2025-07-17T08:42:52.7894648Z Profiles: HSA_PROFILE_BASE 2025-07-17T08:42:52.7894916Z Default Rounding Mode: NEAR 2025-07-17T08:42:52.7895172Z Default Rounding Mode: NEAR 2025-07-17T08:42:52.7895410Z Fast f16: TRUE 2025-07-17T08:42:52.7895648Z Workgroup Max Size: 1024(0x400) 2025-07-17T08:42:52.7895883Z Workgroup Max Size per Dimension: 2025-07-17T08:42:52.7896079Z x 1024(0x400) 2025-07-17T08:42:52.7896290Z y 1024(0x400) 2025-07-17T08:42:52.7896489Z z 1024(0x400) 2025-07-17T08:42:52.7896704Z Grid Max Size: 4294967295(0xffffffff) 2025-07-17T08:42:52.7896926Z Grid Max Size per Dimension: 2025-07-17T08:42:52.7897117Z x 4294967295(0xffffffff) 2025-07-17T08:42:52.7897319Z y 4294967295(0xffffffff) 2025-07-17T08:42:52.7897522Z z 4294967295(0xffffffff) 2025-07-17T08:42:52.7897740Z FBarrier Max Size: 32 2025-07-17T08:42:52.7897957Z ******* 2025-07-17T08:42:52.7898119Z Agent 4 2025-07-17T08:42:52.7898279Z ******* 2025-07-17T08:42:52.7898458Z Name: gfx942 2025-07-17T08:42:52.7898689Z Uuid: GPU-e07d565caf047d31 2025-07-17T08:42:52.7898931Z Marketing Name: AMD Instinct MI300X 2025-07-17T08:42:52.7899177Z Vendor Name: AMD 2025-07-17T08:42:52.7899414Z Feature: KERNEL_DISPATCH 2025-07-17T08:42:52.7899647Z Profile: BASE_PROFILE 2025-07-17T08:42:52.7899893Z Float Round Mode: NEAR 2025-07-17T08:42:52.7900138Z Max Queue Number: 128(0x80) 2025-07-17T08:42:52.7900374Z Queue Min Size: 64(0x40) 2025-07-17T08:42:52.7900605Z Queue Max Size: 131072(0x20000) 2025-07-17T08:42:52.7900840Z Queue Type: MULTI 2025-07-17T08:42:52.7901187Z Node: 3 2025-07-17T08:42:52.7901519Z Device Type: GPU 2025-07-17T08:42:52.7901744Z Cache Info: 2025-07-17T08:42:52.7901927Z L1: 32(0x20) KB 2025-07-17T08:42:52.7902135Z L2: 4096(0x1000) KB 2025-07-17T08:42:52.7902338Z L3: 262144(0x40000) KB 2025-07-17T08:42:52.7902548Z Chip ID: 29857(0x74a1) 2025-07-17T08:42:52.7902781Z ASIC Revision: 1(0x1) 2025-07-17T08:42:52.7903020Z Cacheline Size: 128(0x80) 2025-07-17T08:42:52.7903260Z Max Clock Freq. (MHz): 2100 2025-07-17T08:42:52.7903485Z BDFID: 50944 2025-07-17T08:42:52.7903713Z Internal Node ID: 3 2025-07-17T08:42:52.7903953Z Compute Unit: 304 2025-07-17T08:42:52.7904191Z SIMDs per CU: 4 2025-07-17T08:42:52.7904432Z Shader Engines: 32 2025-07-17T08:42:52.7904674Z Shader Arrs. per Eng.: 1 2025-07-17T08:42:52.7904925Z WatchPts on Addr. Ranges:4 2025-07-17T08:42:52.7905182Z Coherent Host Access: FALSE 2025-07-17T08:42:52.7905408Z Memory Properties: 2025-07-17T08:42:52.7905594Z Features: KERNEL_DISPATCH 2025-07-17T08:42:52.7905823Z Fast F16 Operation: TRUE 2025-07-17T08:42:52.7906064Z Wavefront Size: 64(0x40) 2025-07-17T08:42:52.7906313Z Workgroup Max Size: 1024(0x400) 2025-07-17T08:42:52.7906547Z Workgroup Max Size per Dimension: 2025-07-17T08:42:52.7906743Z x 1024(0x400) 2025-07-17T08:42:52.7906950Z y 1024(0x400) 2025-07-17T08:42:52.7907148Z z 1024(0x400) 2025-07-17T08:42:52.7907370Z Max Waves Per CU: 32(0x20) 2025-07-17T08:42:52.7907608Z Max Work-item Per CU: 2048(0x800) 2025-07-17T08:42:52.7907860Z Grid Max Size: 4294967295(0xffffffff) 2025-07-17T08:42:52.7908078Z Grid Max Size per Dimension: 2025-07-17T08:42:52.7908264Z x 4294967295(0xffffffff) 2025-07-17T08:42:52.7908467Z y 4294967295(0xffffffff) 2025-07-17T08:42:52.7908671Z z 4294967295(0xffffffff) 2025-07-17T08:42:52.7908902Z Max fbarriers/Workgrp: 32 2025-07-17T08:42:52.7909174Z Packet Processor uCode:: 177 2025-07-17T08:42:52.7909431Z SDMA engine uCode:: 24 2025-07-17T08:42:52.7909679Z IOMMU Support:: None 2025-07-17T08:42:52.7909891Z Pool Info: 2025-07-17T08:42:52.7910057Z Pool 1 2025-07-17T08:42:52.7910263Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-07-17T08:42:52.7910494Z Size: 201310208(0xbffc000) KB 2025-07-17T08:42:52.7910745Z Allocatable: TRUE 2025-07-17T08:42:52.7910997Z Alloc Granule: 4KB 2025-07-17T08:42:52.7911249Z Alloc Recommended Granule:2048KB 2025-07-17T08:42:52.7911511Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7911904Z Accessible by all: FALSE 2025-07-17T08:42:52.7912122Z Pool 2 2025-07-17T08:42:52.7912421Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-07-17T08:42:52.7912656Z Size: 201310208(0xbffc000) KB 2025-07-17T08:42:52.7912887Z Allocatable: TRUE 2025-07-17T08:42:52.7913130Z Alloc Granule: 4KB 2025-07-17T08:42:52.7913382Z Alloc Recommended Granule:2048KB 2025-07-17T08:42:52.7913640Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7913967Z Accessible by all: FALSE 2025-07-17T08:42:52.7914240Z Pool 3 2025-07-17T08:42:52.7914435Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-07-17T08:42:52.7914664Z Size: 201310208(0xbffc000) KB 2025-07-17T08:42:52.7914897Z Allocatable: TRUE 2025-07-17T08:42:52.7915142Z Alloc Granule: 4KB 2025-07-17T08:42:52.7915392Z Alloc Recommended Granule:2048KB 2025-07-17T08:42:52.7915643Z Alloc Alignment: 4KB 2025-07-17T08:42:52.7915887Z Accessible by all: FALSE 2025-07-17T08:42:52.7916105Z Pool 4 2025-07-17T08:42:52.7916292Z Segment: GROUP 2025-07-17T08:42:52.7916513Z Size: 64(0x40) KB 2025-07-17T08:42:52.7916741Z Allocatable: FALSE 2025-07-17T08:42:52.7916980Z Alloc Granule: 0KB 2025-07-17T08:42:52.7917229Z Alloc Recommended Granule:0KB 2025-07-17T08:42:52.7917489Z Alloc Alignment: 0KB 2025-07-17T08:42:52.7917738Z Accessible by all: FALSE 2025-07-17T08:42:52.7917953Z ISA Info: 2025-07-17T08:42:52.7918114Z ISA 1 2025-07-17T08:42:52.7918313Z Name: amdgcn-amd-amdhsa--gfx942:sramecc+:xnack- 2025-07-17T08:42:52.7918578Z Machine Models: HSA_MACHINE_MODEL_LARGE 2025-07-17T08:42:52.7918830Z Profiles: HSA_PROFILE_BASE 2025-07-17T08:42:52.7919079Z Default Rounding Mode: NEAR 2025-07-17T08:42:52.7919338Z Default Rounding Mode: NEAR 2025-07-17T08:42:52.7919579Z Fast f16: TRUE 2025-07-17T08:42:52.7919819Z Workgroup Max Size: 1024(0x400) 2025-07-17T08:42:52.7920058Z Workgroup Max Size per Dimension: 2025-07-17T08:42:52.7920264Z x 1024(0x400) 2025-07-17T08:42:52.7920480Z y 1024(0x400) 2025-07-17T08:42:52.7920681Z z 1024(0x400) 2025-07-17T08:42:52.7920900Z Grid Max Size: 4294967295(0xffffffff) 2025-07-17T08:42:52.7921120Z Grid Max Size per Dimension: 2025-07-17T08:42:52.7921312Z x 4294967295(0xffffffff) 2025-07-17T08:42:52.7921512Z y 4294967295(0xffffffff) 2025-07-17T08:42:52.7921714Z z 4294967295(0xffffffff) 2025-07-17T08:42:52.7921944Z FBarrier Max Size: 32 2025-07-17T08:42:52.7922156Z ISA 2 2025-07-17T08:42:52.7922370Z Name: amdgcn-amd-amdhsa--gfx9-4-generic:sramecc+:xnack- 2025-07-17T08:42:52.7955651Z Machine Models: HSA_MACHINE_MODEL_LARGE 2025-07-17T08:42:52.7956034Z Profiles: HSA_PROFILE_BASE 2025-07-17T08:42:52.7956292Z Default Rounding Mode: NEAR 2025-07-17T08:42:52.7956553Z Default Rounding Mode: NEAR 2025-07-17T08:42:52.7956797Z Fast f16: TRUE 2025-07-17T08:42:52.7957038Z Workgroup Max Size: 1024(0x400) 2025-07-17T08:42:52.7957271Z Workgroup Max Size per Dimension: 2025-07-17T08:42:52.7957474Z x 1024(0x400) 2025-07-17T08:42:52.7957689Z y 1024(0x400) 2025-07-17T08:42:52.7957883Z z 1024(0x400) 2025-07-17T08:42:52.7958111Z Grid Max Size: 4294967295(0xffffffff) 2025-07-17T08:42:52.7958334Z Grid Max Size per Dimension: 2025-07-17T08:42:52.7958524Z x 4294967295(0xffffffff) 2025-07-17T08:42:52.7958729Z y 4294967295(0xffffffff) 2025-07-17T08:42:52.7958931Z z 4294967295(0xffffffff) 2025-07-17T08:42:52.7959156Z FBarrier Max Size: 32 2025-07-17T08:42:52.7959371Z *** Done *** 2025-07-17T08:42:52.8005278Z + rocminfo 2025-07-17T08:42:52.8006809Z + grep -E 'Name:.*\sgfx|Marketing' 2025-07-17T08:42:53.0629284Z Marketing Name: AMD EPYC 9534 64-Core Processor 2025-07-17T08:42:53.0629672Z Marketing Name: AMD EPYC 9534 64-Core Processor 2025-07-17T08:42:53.0629941Z Name: gfx942 2025-07-17T08:42:53.0630192Z Marketing Name: AMD Instinct MI300X 2025-07-17T08:42:53.0630487Z Name: gfx942 2025-07-17T08:42:53.0630751Z Marketing Name: AMD Instinct MI300X 2025-07-17T08:42:53.0879183Z + MAYBE_ROCM=rocm/ 2025-07-17T08:42:53.0879404Z + [[ linux-noble-rocm-py3.12-mi300 == *xpu* ]] 2025-07-17T08:42:53.0879691Z + [[ linux-noble-rocm-py3.12-mi300 != *-bazel-* ]] 2025-07-17T08:42:53.0879928Z + pip_install ninja==1.10.2 2025-07-17T08:42:53.0880195Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-07-17T08:42:53.0880501Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-07-17T08:42:54.6483045Z Collecting ninja==1.10.2 2025-07-17T08:42:54.6699472Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl.metadata (5.0 kB) 2025-07-17T08:42:54.6778588Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2025-07-17T08:42:54.8457055Z Installing collected packages: ninja 2025-07-17T08:42:54.8457441Z Attempting uninstall: ninja 2025-07-17T08:42:54.8474802Z Found existing installation: ninja 1.11.1.3 2025-07-17T08:42:54.8496394Z Uninstalling ninja-1.11.1.3: 2025-07-17T08:42:54.8524248Z Successfully uninstalled ninja-1.11.1.3 2025-07-17T08:42:54.8703629Z Successfully installed ninja-1.10.2 2025-07-17T08:42:54.9163785Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/opt/rocm/llvm/bin:/opt/rocm/opencl/bin:/opt/rocm/hip/bin:/opt/rocm/hcc/bin:/opt/rocm/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-07-17T08:42:54.9165158Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/opt/rocm/llvm/bin:/opt/rocm/opencl/bin:/opt/rocm/hip/bin:/opt/rocm/hcc/bin:/opt/rocm/bin:/opt/conda/envs/py_3.12/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-07-17T08:42:54.9166002Z + [[ linux-noble-rocm-py3.12-mi300 == *aarch64* ]] 2025-07-17T08:42:54.9166257Z + [[ linux-noble-rocm-py3.12-mi300 == *asan* ]] 2025-07-17T08:42:54.9167130Z + [[ linux-noble-rocm-py3.12-mi300 == *-debug* ]] 2025-07-17T08:42:54.9167636Z + [[ linux-noble-rocm-py3.12-mi300 != *-bazel-* ]] 2025-07-17T08:42:54.9168052Z + echo 'We are not in debug mode: linux-noble-rocm-py3.12-mi300. Expect the assertion to pass' 2025-07-17T08:42:54.9168508Z We are not in debug mode: linux-noble-rocm-py3.12-mi300. Expect the assertion to pass 2025-07-17T08:42:54.9168802Z + cd test 2025-07-17T08:42:54.9169047Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-07-17T08:42:56.9412478Z + [[ default == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-07-17T08:42:56.9412782Z + [[ default == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-07-17T08:42:56.9413026Z + [[ default == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-07-17T08:42:56.9413287Z + DYNAMO_BENCHMARK_FLAGS=() 2025-07-17T08:42:56.9413477Z + [[ default == *pr_time_benchmarks* ]] 2025-07-17T08:42:56.9413688Z + [[ default == *dynamo_eager* ]] 2025-07-17T08:42:56.9413882Z + [[ default == *aot_eager* ]] 2025-07-17T08:42:56.9414094Z + [[ default == *aot_inductor* ]] 2025-07-17T08:42:56.9414283Z + [[ default == *max_autotune_inductor* ]] 2025-07-17T08:42:56.9414504Z + [[ default == *inductor* ]] 2025-07-17T08:42:56.9414678Z + [[ default == *dynamic* ]] 2025-07-17T08:42:56.9414852Z + [[ default == *cpu* ]] 2025-07-17T08:42:56.9415055Z + DYNAMO_BENCHMARK_FLAGS+=(--device cuda) 2025-07-17T08:42:56.9430634Z + [[ linux-noble-rocm-py3.12-mi300 == *libtorch* ]] 2025-07-17T08:42:56.9430930Z + [[ linux-noble-rocm-py3.12-mi300 == *-bazel-* ]] 2025-07-17T08:42:56.9435057Z + cd test 2025-07-17T08:42:56.9437012Z + python -c 'import torch; print(torch.__config__.show())' 2025-07-17T08:42:58.2075653Z PyTorch built with: 2025-07-17T08:42:58.2075914Z - GCC 11.4 2025-07-17T08:42:58.2076078Z - C++ Version: 201703 2025-07-17T08:42:58.2076460Z - Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-07-17T08:42:58.2076911Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-07-17T08:42:58.2077216Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-07-17T08:42:58.2077444Z - LAPACK is enabled (usually provided by MKL) 2025-07-17T08:42:58.2077654Z - NNPACK is enabled 2025-07-17T08:42:58.2077825Z - CPU capability usage: AVX512 2025-07-17T08:42:58.2078006Z - HIP Runtime 6.4.43483 2025-07-17T08:42:58.2078174Z - MIOpen 3.4.0 2025-07-17T08:42:58.2078321Z - Magma 2.7.2 2025-07-17T08:42:58.2080925Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, COMMIT_SHA=a38f433be2e94a64b095a44ba39879d02d0c2316, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DLIBKINETO_NOXPUPTI=ON -DUSE_FBGEMM -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, TORCH_VERSION=2.9.0, USE_CUDA=OFF, USE_CUDNN=OFF, USE_CUSPARSELT=OFF, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=ON, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-07-17T08:42:58.2083543Z 2025-07-17T08:42:58.4934504Z + cd test 2025-07-17T08:42:58.4934834Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-07-17T08:42:59.5644608Z ATen/Parallel: 2025-07-17T08:42:59.5644905Z at::get_num_threads() : 128 2025-07-17T08:42:59.5645124Z at::get_num_interop_threads() : 128 2025-07-17T08:42:59.5645344Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-07-17T08:42:59.5646166Z omp_get_max_threads() : 128 2025-07-17T08:42:59.5646748Z Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-07-17T08:42:59.5647122Z mkl_get_max_threads() : 128 2025-07-17T08:42:59.5647382Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-07-17T08:42:59.5647668Z std::thread::hardware_concurrency() : 256 2025-07-17T08:42:59.5647878Z Environment variables: 2025-07-17T08:42:59.5648049Z OMP_NUM_THREADS : [not set] 2025-07-17T08:42:59.5648222Z MKL_NUM_THREADS : [not set] 2025-07-17T08:42:59.5648403Z ATen parallel backend: OpenMP 2025-07-17T08:42:59.5648525Z 2025-07-17T08:43:00.0298430Z + [[ default == *numpy_2* ]] 2025-07-17T08:43:00.0298759Z + [[ linux-noble-rocm-py3.12-mi300 == *aarch64* ]] 2025-07-17T08:43:00.0299002Z + [[ default == *backward* ]] 2025-07-17T08:43:00.0299185Z + [[ default == *xla* ]] 2025-07-17T08:43:00.0299356Z + [[ default == *executorch* ]] 2025-07-17T08:43:00.0299552Z + [[ default == \j\i\t\_\l\e\g\a\c\y ]] 2025-07-17T08:43:00.0299806Z + [[ linux-noble-rocm-py3.12-mi300 == *libtorch* ]] 2025-07-17T08:43:00.0300031Z + [[ default == distributed ]] 2025-07-17T08:43:00.0300229Z + [[ default == *operator_benchmark* ]] 2025-07-17T08:43:00.0300459Z + [[ default == *inductor_distributed* ]] 2025-07-17T08:43:00.0300659Z + [[ default == *inductor-halide* ]] 2025-07-17T08:43:00.0300864Z + [[ default == *inductor-triton-cpu* ]] 2025-07-17T08:43:00.0301090Z + [[ default == *inductor-micro-benchmark* ]] 2025-07-17T08:43:00.0301303Z + [[ default == *huggingface* ]] 2025-07-17T08:43:00.0301492Z + [[ default == *timm* ]] 2025-07-17T08:43:00.0301666Z + [[ default == cachebench ]] 2025-07-17T08:43:00.0301859Z + [[ default == verify_cachebench ]] 2025-07-17T08:43:00.0302057Z + [[ default == *torchbench* ]] 2025-07-17T08:43:00.0302245Z + [[ default == *inductor_cpp_wrapper* ]] 2025-07-17T08:43:00.0302440Z + [[ default == *inductor* ]] 2025-07-17T08:43:00.0302615Z + [[ default == *einops* ]] 2025-07-17T08:43:00.0302801Z + [[ default == *dynamo_wrapped* ]] 2025-07-17T08:43:00.0303015Z + [[ linux-noble-rocm-py3.12-mi300 == *rocm* ]] 2025-07-17T08:43:00.0303228Z + [[ -n '' ]] 2025-07-17T08:43:00.0303382Z + [[ 2 == 1 ]] 2025-07-17T08:43:00.0303532Z + [[ 2 == 2 ]] 2025-07-17T08:43:00.0303671Z + [[ 6 -gt 1 ]] 2025-07-17T08:43:00.0303821Z + install_torchvision 2025-07-17T08:43:00.0303988Z + local orig_preload 2025-07-17T08:43:00.0304142Z + local commit 2025-07-17T08:43:00.0304317Z ++ get_pinned_commit vision 2025-07-17T08:43:00.0304506Z ++ cat .github/ci_commit_pins/vision.txt 2025-07-17T08:43:00.0320499Z + commit=966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-07-17T08:43:00.0320770Z + orig_preload= 2025-07-17T08:43:00.0320935Z + '[' -n '' ']' 2025-07-17T08:43:00.0321302Z + pip_install --no-use-pep517 git+https://github.com/pytorch/vision.git@966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-07-17T08:43:00.0321732Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-07-17T08:43:00.0322198Z + python3 -m pip install --progress-bar off --no-use-pep517 git+https://github.com/pytorch/vision.git@966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-07-17T08:43:00.2604587Z Collecting git+https://github.com/pytorch/vision.git@966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-07-17T08:43:00.2606325Z Cloning https://github.com/pytorch/vision.git (to revision 966da7e46f65d6d49df3e31214470a4fe5cc8e66) to /tmp/pip-req-build-0dsdkg_a 2025-07-17T08:43:00.2654897Z Running command git clone --filter=blob:none --quiet https://github.com/pytorch/vision.git /tmp/pip-req-build-0dsdkg_a 2025-07-17T08:43:02.6075961Z Running command git rev-parse -q --verify 'sha^966da7e46f65d6d49df3e31214470a4fe5cc8e66' 2025-07-17T08:43:02.6098642Z Running command git fetch -q https://github.com/pytorch/vision.git 966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-07-17T08:43:02.9284816Z Running command git checkout -q 966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-07-17T08:43:03.4011363Z Resolved https://github.com/pytorch/vision.git to commit 966da7e46f65d6d49df3e31214470a4fe5cc8e66 2025-07-17T08:43:05.6827923Z Preparing metadata (setup.py) ... [?25l- \ | / done 2025-07-17T08:43:05.6861263Z [?25hRequirement already satisfied: numpy in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.22.0a0+966da7e) (1.26.2) 2025-07-17T08:43:05.6862653Z Requirement already satisfied: torch in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.22.0a0+966da7e) (2.9.0a0+gita38f433) 2025-07-17T08:43:05.6865184Z Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torchvision==0.22.0a0+966da7e) (11.0.0) 2025-07-17T08:43:05.6910107Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (3.18.0) 2025-07-17T08:43:05.6912862Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (4.14.1) 2025-07-17T08:43:05.6921483Z Requirement already satisfied: setuptools in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (80.9.0) 2025-07-17T08:43:05.6924072Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (1.13.3) 2025-07-17T08:43:05.6926166Z Requirement already satisfied: networkx in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (2.8.8) 2025-07-17T08:43:05.6928034Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (3.1.6) 2025-07-17T08:43:05.6930535Z Requirement already satisfied: fsspec in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from torch->torchvision==0.22.0a0+966da7e) (2025.5.1) 2025-07-17T08:43:05.6940619Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from sympy>=1.13.3->torch->torchvision==0.22.0a0+966da7e) (1.3.0) 2025-07-17T08:43:05.7023226Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.12/lib/python3.12/site-packages (from jinja2->torch->torchvision==0.22.0a0+966da7e) (3.0.2) 2025-07-17T08:43:05.7082028Z Building wheels for collected packages: torchvision 2025-07-17T08:43:05.7154161Z  DEPRECATION: Building 'torchvision' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'torchvision'. Discussion can be found at https://github.com/pypa/pip/issues/6334 2025-07-17T08:43:52.0006360Z  Building wheel for torchvision (setup.py) ... [?25l- \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ done 2025-07-17T08:43:52.0028794Z [?25h Created wheel for torchvision: filename=torchvision-0.22.0a0+966da7e-cp312-cp312-linux_x86_64.whl size=1571720 sha256=5bf8ba486027bd067bd1f2cf5f0a33c5daf8cc767bbca089c7a3218703a90043 2025-07-17T08:43:52.0029598Z Stored in directory: /var/lib/jenkins/.cache/pip/wheels/10/ba/61/eb5228b3631dc6bb4f478b3aa59575551a5473e4596e4c001a 2025-07-17T08:43:52.0060269Z Successfully built torchvision 2025-07-17T08:43:52.1519113Z Installing collected packages: torchvision 2025-07-17T08:43:52.5160323Z Successfully installed torchvision-0.22.0a0+966da7e 2025-07-17T08:43:52.6689332Z + '[' -n '' ']' 2025-07-17T08:43:52.6689573Z + test_python_shard 2 2025-07-17T08:43:52.6689759Z + [[ -z 6 ]] 2025-07-17T08:43:52.6690183Z + python test/run_test.py --exclude-jit-executor --exclude-distributed-tests --shard 2 6 --verbose --upload-artifacts-while-running 2025-07-17T08:43:55.2270342Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T08:43:55.2271451Z import pkg_resources 2025-07-17T08:43:55.7420262Z Excluding test_cuda_nvml_based_avail on ROCm 2025-07-17T08:43:55.7967313Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to /var/lib/jenkins/pytorch/test/.pytorch-disabled-tests.json 2025-07-17T08:43:58.8311848Z Ignoring disabled issues: [''] 2025-07-17T08:43:58.8515917Z Found test times from artifacts 2025-07-17T08:43:58.9302087Z Found test times from artifacts 2025-07-17T08:43:58.9336413Z Running all tests 2025-07-17T08:43:58.9592256Z Running parallel tests on 2 processes 2025-07-17T08:43:58.9595169Z Name: tests to run (est. time: 106.83min) 2025-07-17T08:43:58.9595440Z Serial tests (0): 2025-07-17T08:43:58.9595620Z Parallel tests (85): 2025-07-17T08:43:58.9595830Z dynamo/test_inline_and_install 1/1 2025-07-17T08:43:58.9596054Z functorch/test_ops 3/5 2025-07-17T08:43:58.9596293Z inductor/test_torchinductor_codegen_dynamic_shapes 1/5 2025-07-17T08:43:58.9596542Z test_nestedtensor 2/3 2025-07-17T08:43:58.9596751Z inductor/test_torchinductor_dynamic_shapes 2/6 2025-07-17T08:43:58.9596981Z test_transformers 1/1 2025-07-17T08:43:58.9597154Z test_decomp 2/12 2025-07-17T08:43:58.9598847Z test_decomp 10/12 2025-07-17T08:43:58.9599050Z test_decomp 12/12 2025-07-17T08:43:58.9599225Z test_foreach 1/1 2025-07-17T08:43:58.9599408Z torch_np/numpy_tests/linalg/test_linalg 1/1 2025-07-17T08:43:58.9599616Z test_jit_fuser_te 1/1 2025-07-17T08:43:58.9599779Z test_mkldnn 1/1 2025-07-17T08:43:58.9599950Z inductor/test_torchinductor 2/2 2025-07-17T08:43:58.9600138Z test_autoload 1/1 2025-07-17T08:43:58.9600296Z test_jit_llga_fuser 1/1 2025-07-17T08:43:58.9600473Z dynamo/test_cudagraphs 1/1 2025-07-17T08:43:58.9600688Z dynamo/test_cudagraphs_expandable_segments 1/1 2025-07-17T08:43:58.9600903Z dynamo/test_debug_utils 1/1 2025-07-17T08:43:58.9601086Z dynamo/test_decorators 1/1 2025-07-17T08:43:58.9601278Z dynamo/test_deque_reconstruct 1/1 2025-07-17T08:43:58.9601480Z dynamo/test_deviceguard 1/1 2025-07-17T08:43:58.9601662Z dynamo/test_dicts 1/1 2025-07-17T08:43:58.9601840Z dynamo/test_dynamic_shapes 1/3 2025-07-17T08:43:58.9602037Z export/test_serialize 1/1 2025-07-17T08:43:58.9602208Z export/test_sparse 1/1 2025-07-17T08:43:58.9602389Z export/test_swap 1/1 2025-07-17T08:43:58.9602565Z export/test_tools 1/1 2025-07-17T08:43:58.9602731Z export/test_torchbind 1/1 2025-07-17T08:43:58.9602904Z export/test_tree_utils 1/1 2025-07-17T08:43:58.9603079Z export/test_unflatten 1/1 2025-07-17T08:43:58.9603270Z export/test_unflatten_training_ir 1/1 2025-07-17T08:43:58.9603472Z export/test_upgrader 1/1 2025-07-17T08:43:58.9603646Z export/test_verifier 1/1 2025-07-17T08:43:58.9603831Z functorch/test_ac_knapsack 1/1 2025-07-17T08:43:58.9604014Z functorch/test_ac_logging 1/1 2025-07-17T08:43:58.9604208Z functorch/test_control_flow 1/1 2025-07-17T08:43:58.9604420Z inductor/test_aot_inductor_custom_ops 1/1 2025-07-17T08:43:58.9604639Z inductor/test_aot_inductor_package 1/1 2025-07-17T08:43:58.9604844Z inductor/test_b2b_gemm 1/1 2025-07-17T08:43:58.9605034Z inductor/test_compile_subprocess 2/2 2025-07-17T08:43:58.9605251Z inductor/test_compiled_optimizers 2/3 2025-07-17T08:43:58.9605461Z inductor/test_cuda_repro 1/1 2025-07-17T08:43:58.9605656Z inductor/test_cudagraph_trees 1/1 2025-07-17T08:43:58.9605894Z inductor/test_cudagraph_trees_expandable_segments 1/1 2025-07-17T08:43:58.9606148Z inductor/test_custom_post_grad_passes 1/1 2025-07-17T08:43:58.9606356Z inductor/test_cutlass_evt 1/1 2025-07-17T08:43:58.9606549Z inductor/test_debug_trace 1/1 2025-07-17T08:43:58.9606754Z inductor/test_decompose_mem_bound_mm 1/1 2025-07-17T08:43:58.9607437Z inductor/test_dependencies 1/1 2025-07-17T08:43:58.9607644Z inductor/test_efficient_conv_bn_eval 1/1 2025-07-17T08:43:58.9608054Z inductor/test_extension_backend 1/1 2025-07-17T08:43:58.9608270Z inductor/test_external_callables 1/1 2025-07-17T08:43:58.9608484Z inductor/test_flex_decoding 1/3 2025-07-17T08:43:58.9608681Z inductor/test_mps_basic 1/1 2025-07-17T08:43:58.9608881Z inductor/test_ordered_set 1/1 2025-07-17T08:43:58.9609072Z inductor/test_pad_mm 1/1 2025-07-17T08:43:58.9609253Z inductor/test_padding 1/1 2025-07-17T08:43:58.9609443Z inductor/test_pattern_matcher 1/1 2025-07-17T08:43:58.9609636Z inductor/test_perf 1/1 2025-07-17T08:43:58.9609813Z inductor/test_profiler 1/1 2025-07-17T08:43:58.9609999Z inductor/test_quantization 1/1 2025-07-17T08:43:58.9610204Z inductor/test_scatter_optimization 1/1 2025-07-17T08:43:58.9610420Z inductor/test_select_algorithm 1/1 2025-07-17T08:43:58.9610623Z inductor/test_snode_runtime 1/1 2025-07-17T08:43:58.9610840Z inductor/test_split_cat_fx_aten_passes 1/1 2025-07-17T08:43:58.9611061Z inductor/test_split_cat_fx_passes 1/1 2025-07-17T08:43:58.9611276Z inductor/test_static_cuda_launcher 1/1 2025-07-17T08:43:58.9611473Z inductor/test_subgraph_choice 1/1 2025-07-17T08:43:58.9611659Z inductor/test_torchbind 1/1 2025-07-17T08:43:58.9611885Z inductor/test_torchinductor_codegen_config_overrides 1/1 2025-07-17T08:43:58.9612143Z inductor/test_torchinductor_opinfo 1/13 2025-07-17T08:43:58.9612363Z inductor/test_torchinductor_opinfo 8/13 2025-07-17T08:43:58.9612567Z inductor/test_torchinductor_opinfo 11/13 2025-07-17T08:43:58.9612770Z optim/test_swa_utils 1/1 2025-07-17T08:43:58.9612968Z test_out_dtype_op 1/1 2025-07-17T08:43:58.9613135Z test_package 1/1 2025-07-17T08:43:58.9613297Z test_per_overload_api 1/1 2025-07-17T08:43:58.9613471Z test_prims 1/1 2025-07-17T08:43:58.9613626Z test_proxy_tensor 1/1 2025-07-17T08:43:58.9613789Z test_pruning_op 1/1 2025-07-17T08:43:58.9613958Z test_public_bindings 1/1 2025-07-17T08:43:58.9614125Z test_pytree 1/1 2025-07-17T08:43:58.9622340Z test_quantization 1/3 2025-07-17T08:43:58.9622568Z test_quantization 3/3 2025-07-17T08:43:58.9622810Z Name: excluded (est. time: 0.0min) 2025-07-17T08:43:58.9623011Z Serial tests (0): 2025-07-17T08:43:58.9623165Z Parallel tests (0): 2025-07-17T08:43:58.9623422Z Running dynamo/test_inline_and_install 1/1 ... [2025-07-17 08:43:58.960357] 2025-07-17T08:43:58.9623703Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:43:58.9624364Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_inline_and_install.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:43:58.960671] 2025-07-17T08:44:03.0836507Z 2025-07-17T08:44:03.0837405Z dynamo/test_inline_and_install 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_inline_and_install_1.1_4c02cf0438f854e2_.log 2025-07-17T08:44:03.0837986Z Running 0 items in this shard: 2025-07-17T08:44:03.0838127Z 2025-07-17T08:44:03.0838292Z GITHUB_RUN_ID, GITHUB_RUN_ATTEMPT, or ARTIFACTS_FILE_SUFFIX not set, not uploading 2025-07-17T08:44:03.0838593Z Uploading artifacts took 0.00 seconds 2025-07-17T08:44:03.0838893Z Running functorch/test_ops 3/5 ... [2025-07-17 08:44:03.083728] 2025-07-17T08:44:03.0839349Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:44:03.0841690Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'functorch/test_ops.py', '-m', 'serial', '--shard-id=3', '--num-shards=5', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:44:03.084019] 2025-07-17T08:44:09.6163259Z 2025-07-17T08:44:09.6164061Z functorch/test_ops 3/5 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ops_3.5_17200f90d6b0db27_.log 2025-07-17T08:44:09.6165149Z Running 0 items in this shard: 2025-07-17T08:44:09.6165281Z 2025-07-17T08:44:09.6165735Z Running inductor/test_torchinductor_codegen_dynamic_shapes 1/5 ... [2025-07-17 08:44:09.616381] 2025-07-17T08:44:09.6166083Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:44:09.6168647Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor_codegen_dynamic_shapes.py', '-m', 'serial', '--shard-id=1', '--num-shards=5', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:44:09.616669] 2025-07-17T08:44:29.1759793Z 2025-07-17T08:44:29.1760791Z inductor/test_torchinductor_codegen_dynamic_shapes 1/5 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_codegen_dynamic_shapes_1.5_ceb5486a4671a354_.log 2025-07-17T08:44:29.1761782Z Running 1 items in this shard: test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_large_block_sizes_dynamic_shapes_cuda 2025-07-17T08:44:29.1762240Z 2025-07-17T08:44:29.1766487Z Running test_nestedtensor 2/3 ... [2025-07-17 08:44:29.176011] 2025-07-17T08:44:29.1766792Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:44:29.1767444Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_nestedtensor.py', '-m', 'serial', '--shard-id=2', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:44:29.176312] 2025-07-17T08:44:34.0510956Z 2025-07-17T08:44:34.0511787Z test_nestedtensor 2/3 was successful, full logs can be found in artifacts with path test/test-reports/test_nestedtensor_2.3_6fe386f873be5ded_.log 2025-07-17T08:44:34.0512289Z Running 0 items in this shard: 2025-07-17T08:44:34.0512417Z 2025-07-17T08:44:34.0512633Z Running inductor/test_torchinductor_dynamic_shapes 2/6 ... [2025-07-17 08:44:34.051111] 2025-07-17T08:44:34.0512997Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:44:34.0516138Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor_dynamic_shapes.py', '-m', 'serial', '--shard-id=2', '--num-shards=6', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:44:34.051420] 2025-07-17T08:44:48.9972978Z 2025-07-17T08:44:48.9973953Z inductor/test_torchinductor_dynamic_shapes 2/6 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_dynamic_shapes_2.6_d46e6c7339cbb571_.log 2025-07-17T08:44:48.9974902Z Running 1 items in this shard: test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_wrapper_codegen_statically_known_int_or_none_cuda 2025-07-17T08:44:48.9975348Z 2025-07-17T08:44:48.9975470Z Running test_transformers 1/1 ... [2025-07-17 08:44:48.997271] 2025-07-17T08:44:48.9975753Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:44:48.9976932Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_transformers.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:44:48.997567] 2025-07-17T08:44:55.3262038Z 2025-07-17T08:44:55.3262770Z test_transformers 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_transformers_1.1_94693c4dc6de8225_.log 2025-07-17T08:44:55.3263241Z Running 0 items in this shard: 2025-07-17T08:44:55.3263382Z 2025-07-17T08:44:55.3263797Z Running test_decomp 2/12 ... [2025-07-17 08:44:55.326194] 2025-07-17T08:44:55.3264088Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:44:55.3266116Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_decomp.py', '-m', 'serial', '--shard-id=2', '--num-shards=12', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:44:55.326472] 2025-07-17T08:45:01.1552554Z 2025-07-17T08:45:01.1554293Z test_decomp 2/12 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_2.12_3e3debfdfe7ffd32_.log 2025-07-17T08:45:01.1554790Z Running 0 items in this shard: 2025-07-17T08:45:01.1554919Z 2025-07-17T08:45:01.1555025Z Running test_decomp 10/12 ... [2025-07-17 08:45:01.155248] 2025-07-17T08:45:01.1555271Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:45:01.1556915Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_decomp.py', '-m', 'serial', '--shard-id=10', '--num-shards=12', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:45:01.155548] 2025-07-17T08:45:07.0816295Z 2025-07-17T08:45:07.0816946Z test_decomp 10/12 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_10.12_427ea4e4dfd4ad0d_.log 2025-07-17T08:45:07.0817437Z Running 0 items in this shard: 2025-07-17T08:45:07.0817561Z 2025-07-17T08:45:07.0818259Z Running test_decomp 12/12 ... [2025-07-17 08:45:07.081669] 2025-07-17T08:45:07.0818520Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:45:07.0821207Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_decomp.py', '-m', 'serial', '--shard-id=12', '--num-shards=12', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:45:07.081957] 2025-07-17T08:45:13.0083132Z 2025-07-17T08:45:13.0083930Z test_decomp 12/12 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_12.12_d3952277693feb5d_.log 2025-07-17T08:45:13.0084396Z Running 0 items in this shard: 2025-07-17T08:45:13.0084524Z 2025-07-17T08:45:13.0084656Z Running test_foreach 1/1 ... [2025-07-17 08:45:13.008354] 2025-07-17T08:45:13.0084904Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:45:13.0087896Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_foreach.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:45:13.008658] 2025-07-17T08:45:22.0661703Z 2025-07-17T08:45:22.0662412Z test_foreach 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_foreach_1.1_d8c13376cbf30e57_.log 2025-07-17T08:45:22.0662881Z Running 0 items in this shard: 2025-07-17T08:45:22.0663008Z 2025-07-17T08:45:22.0663175Z Running torch_np/numpy_tests/linalg/test_linalg 1/1 ... [2025-07-17 08:45:22.066172] 2025-07-17T08:45:22.0663478Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:45:22.0665710Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'torch_np/numpy_tests/linalg/test_linalg.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:45:22.066439] 2025-07-17T08:45:25.7280217Z 2025-07-17T08:45:25.7281103Z torch_np/numpy_tests/linalg/test_linalg 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.linalg.test_linalg_1.1_0ae78d2dd3f4d0d9_.log 2025-07-17T08:45:25.7281680Z Running 0 items in this shard: 2025-07-17T08:45:25.7281815Z 2025-07-17T08:45:25.7282168Z Running test_jit_fuser_te 1/1 ... [2025-07-17 08:45:25.728118] 2025-07-17T08:45:25.7282440Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:45:25.7285144Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_jit_fuser_te.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:45:25.728391] 2025-07-17T08:45:32.7317651Z 2025-07-17T08:45:32.7318294Z test_jit_fuser_te 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_jit_fuser_te_1.1_b5afe02d1163a70f_.log 2025-07-17T08:45:32.7319393Z Running 0 items in this shard: 2025-07-17T08:45:32.7319774Z 2025-07-17T08:45:32.7319892Z Running test_mkldnn 1/1 ... [2025-07-17 08:45:32.731729] 2025-07-17T08:45:32.7320141Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:45:32.7322171Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_mkldnn.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:45:32.732044] 2025-07-17T08:45:36.5174350Z 2025-07-17T08:45:36.5175103Z test_mkldnn 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_mkldnn_1.1_ecb222eba54317e4_.log 2025-07-17T08:45:36.5175553Z Running 0 items in this shard: 2025-07-17T08:45:36.5175681Z 2025-07-17T08:45:36.5175852Z Running inductor/test_torchinductor 2/2 ... [2025-07-17 08:45:36.517432] 2025-07-17T08:45:36.5176168Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:45:36.5178727Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor.py', '-m', 'serial', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:45:36.517728] 2025-07-17T08:45:44.5791369Z 2025-07-17T08:45:44.5792216Z inductor/test_torchinductor 2/2 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_2.2_b0238e5cee8c5253_.log 2025-07-17T08:45:44.5792752Z Running 0 items in this shard: 2025-07-17T08:45:44.5792883Z 2025-07-17T08:45:44.5793010Z Running test_autoload 1/1 ... [2025-07-17 08:45:44.579084] 2025-07-17T08:45:44.5793256Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:45:44.5795526Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_autoload.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:45:44.579402] 2025-07-17T08:45:47.7116641Z 2025-07-17T08:45:47.7117349Z test_autoload 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_autoload_1.1_b809d1cd69648bac_.log 2025-07-17T08:45:47.7117822Z Running 0 items in this shard: 2025-07-17T08:45:47.7117953Z 2025-07-17T08:45:47.7118070Z Running test_jit_llga_fuser 1/1 ... [2025-07-17 08:45:47.711638] 2025-07-17T08:45:47.7118344Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:45:47.7121146Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_jit_llga_fuser.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:45:47.711939] 2025-07-17T08:45:51.3344667Z 2025-07-17T08:45:51.3345343Z test_jit_llga_fuser 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_jit_llga_fuser_1.1_1b7c3e80cc07f904_.log 2025-07-17T08:45:51.3345855Z Running 0 items in this shard: 2025-07-17T08:45:51.3345980Z 2025-07-17T08:45:51.3347057Z Running dynamo/test_cudagraphs 1/1 ... [2025-07-17 08:45:51.334514] 2025-07-17T08:45:51.3347352Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:45:51.3349636Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_cudagraphs.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:45:51.334826] 2025-07-17T08:45:54.5567432Z 2025-07-17T08:45:54.5568203Z dynamo/test_cudagraphs 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_cudagraphs_1.1_a6f39e08541fe42b_.log 2025-07-17T08:45:54.5568730Z Running 0 items in this shard: 2025-07-17T08:45:54.5568855Z 2025-07-17T08:45:54.5569938Z Running dynamo/test_cudagraphs_expandable_segments 1/1 ... [2025-07-17 08:45:54.556677] 2025-07-17T08:45:54.5570527Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:45:54.5571272Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_cudagraphs_expandable_segments.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:45:54.556989] 2025-07-17T08:45:58.0795370Z 2025-07-17T08:45:58.0796231Z dynamo/test_cudagraphs_expandable_segments 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_cudagraphs_expandable_segments_1.1_3325f34041b4935c_.log 2025-07-17T08:45:58.0796819Z Running 0 items in this shard: 2025-07-17T08:45:58.0796944Z 2025-07-17T08:45:58.0797097Z Running dynamo/test_debug_utils 1/1 ... [2025-07-17 08:45:58.079447] 2025-07-17T08:45:58.0797369Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:45:58.0799049Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_debug_utils.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:45:58.079751] 2025-07-17T08:46:05.4106055Z 2025-07-17T08:46:05.4106836Z dynamo/test_debug_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_debug_utils_1.1_672fdd4cebc93102_.log 2025-07-17T08:46:05.4107337Z Running 0 items in this shard: 2025-07-17T08:46:05.4107464Z 2025-07-17T08:46:05.4107614Z Running dynamo/test_decorators 1/1 ... [2025-07-17 08:46:05.410569] 2025-07-17T08:46:05.4107884Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:05.4110576Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_decorators.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:05.410875] 2025-07-17T08:46:08.6326873Z 2025-07-17T08:46:08.6327699Z dynamo/test_decorators 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_decorators_1.1_6e3d1368cf641b02_.log 2025-07-17T08:46:08.6328224Z Running 0 items in this shard: 2025-07-17T08:46:08.6328352Z 2025-07-17T08:46:08.6328725Z Running dynamo/test_deque_reconstruct 1/1 ... [2025-07-17 08:46:08.632669] 2025-07-17T08:46:08.6329056Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:08.6331071Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_deque_reconstruct.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:08.632967] 2025-07-17T08:46:11.6543291Z 2025-07-17T08:46:11.6544161Z dynamo/test_deque_reconstruct 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_deque_reconstruct_1.1_7a7fdd61230f030b_.log 2025-07-17T08:46:11.6544726Z Running 0 items in this shard: 2025-07-17T08:46:11.6544859Z 2025-07-17T08:46:11.6545644Z Running dynamo/test_deviceguard 1/1 ... [2025-07-17 08:46:11.654422] 2025-07-17T08:46:11.6545932Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:11.6549982Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_deviceguard.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:11.654796] 2025-07-17T08:46:15.3279138Z 2025-07-17T08:46:15.3280394Z dynamo/test_deviceguard 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_deviceguard_1.1_9930752c9c8e3ce2_.log 2025-07-17T08:46:15.3281207Z Running 0 items in this shard: 2025-07-17T08:46:15.3281399Z 2025-07-17T08:46:15.3282314Z Running dynamo/test_dicts 1/1 ... [2025-07-17 08:46:15.327895] 2025-07-17T08:46:15.3282724Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:15.3284535Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_dicts.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:15.328205] 2025-07-17T08:46:18.5998683Z 2025-07-17T08:46:18.5999607Z dynamo/test_dicts 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_dicts_1.1_6fb398355e47f43d_.log 2025-07-17T08:46:18.6000088Z Running 0 items in this shard: 2025-07-17T08:46:18.6000219Z 2025-07-17T08:46:18.6000363Z Running dynamo/test_dynamic_shapes 1/3 ... [2025-07-17 08:46:18.599871] 2025-07-17T08:46:18.6000647Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:18.6003426Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_dynamic_shapes.py', '-m', 'serial', '--shard-id=1', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:18.600174] 2025-07-17T08:46:27.5855369Z 2025-07-17T08:46:27.5856208Z dynamo/test_dynamic_shapes 1/3 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_dynamic_shapes_1.3_7475803d0f09c233_.log 2025-07-17T08:46:27.5856747Z Running 0 items in this shard: 2025-07-17T08:46:27.5856879Z 2025-07-17T08:46:27.5857008Z Running export/test_serialize 1/1 ... [2025-07-17 08:46:27.585478] 2025-07-17T08:46:27.5857281Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:27.5861041Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_serialize.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:27.585781] 2025-07-17T08:46:30.8578776Z 2025-07-17T08:46:30.8579581Z export/test_serialize 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_serialize_1.1_f7e362af08e4822e_.log 2025-07-17T08:46:30.8580087Z Running 0 items in this shard: 2025-07-17T08:46:30.8580234Z 2025-07-17T08:46:30.8588113Z Running export/test_sparse 1/1 ... [2025-07-17 08:46:30.857934] 2025-07-17T08:46:30.8588422Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:30.8589090Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_sparse.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:30.858261] 2025-07-17T08:46:33.9298067Z 2025-07-17T08:46:33.9298801Z export/test_sparse 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_sparse_1.1_3bec8c4904da597b_.log 2025-07-17T08:46:33.9299285Z Running 0 items in this shard: 2025-07-17T08:46:33.9299447Z 2025-07-17T08:46:33.9300201Z Running export/test_swap 1/1 ... [2025-07-17 08:46:33.929786] 2025-07-17T08:46:33.9300486Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:33.9303000Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_swap.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:33.930104] 2025-07-17T08:46:37.0515788Z 2025-07-17T08:46:37.0516708Z export/test_swap 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_swap_1.1_0ba6e4ff701c1c9e_.log 2025-07-17T08:46:37.0517187Z Running 0 items in this shard: 2025-07-17T08:46:37.0517321Z 2025-07-17T08:46:37.0517775Z Running export/test_tools 1/1 ... [2025-07-17 08:46:37.051622] 2025-07-17T08:46:37.0518040Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:37.0523725Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_tools.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:37.051937] 2025-07-17T08:46:40.2736519Z 2025-07-17T08:46:40.2737269Z export/test_tools 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_tools_1.1_69bc44239dde6b61_.log 2025-07-17T08:46:40.2737736Z Running 0 items in this shard: 2025-07-17T08:46:40.2737882Z 2025-07-17T08:46:40.2738018Z Running export/test_torchbind 1/1 ... [2025-07-17 08:46:40.273600] 2025-07-17T08:46:40.2738292Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:40.2742338Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_torchbind.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:40.273921] 2025-07-17T08:46:43.4954617Z 2025-07-17T08:46:43.4955431Z export/test_torchbind 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_torchbind_1.1_9ff8a32866513af6_.log 2025-07-17T08:46:43.4955927Z Running 0 items in this shard: 2025-07-17T08:46:43.4956055Z 2025-07-17T08:46:43.4956305Z Running export/test_tree_utils 1/1 ... [2025-07-17 08:46:43.495508] 2025-07-17T08:46:43.4956582Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:43.4959746Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_tree_utils.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:43.495806] 2025-07-17T08:46:46.6173398Z 2025-07-17T08:46:46.6174116Z export/test_tree_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_tree_utils_1.1_93a2add606a7ea25_.log 2025-07-17T08:46:46.6174644Z Running 0 items in this shard: 2025-07-17T08:46:46.6174766Z 2025-07-17T08:46:46.6175772Z Running export/test_unflatten 1/1 ... [2025-07-17 08:46:46.617433] 2025-07-17T08:46:46.6176047Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:46.6179197Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_unflatten.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:46.617734] 2025-07-17T08:46:49.6388931Z 2025-07-17T08:46:49.6390103Z export/test_unflatten 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_unflatten_1.1_cfd33a0a2ff41d42_.log 2025-07-17T08:46:49.6390625Z Running 0 items in this shard: 2025-07-17T08:46:49.6390752Z 2025-07-17T08:46:49.6390909Z Running export/test_unflatten_training_ir 1/1 ... [2025-07-17 08:46:49.638891] 2025-07-17T08:46:49.6391235Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:49.6393621Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_unflatten_training_ir.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:49.639183] 2025-07-17T08:46:53.0112454Z 2025-07-17T08:46:53.0113216Z export/test_unflatten_training_ir 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_unflatten_training_ir_1.1_289c6d29647ea337_.log 2025-07-17T08:46:53.0113761Z Running 0 items in this shard: 2025-07-17T08:46:53.0113964Z 2025-07-17T08:46:53.0114136Z Running export/test_upgrader 1/1 ... [2025-07-17 08:46:53.011277] 2025-07-17T08:46:53.0114411Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:53.0117325Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_upgrader.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:53.011578] 2025-07-17T08:46:56.0327619Z 2025-07-17T08:46:56.0328322Z export/test_upgrader 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_upgrader_1.1_6860500b255d4fb5_.log 2025-07-17T08:46:56.0328832Z Running 0 items in this shard: 2025-07-17T08:46:56.0328961Z 2025-07-17T08:46:56.0329098Z Running export/test_verifier 1/1 ... [2025-07-17 08:46:56.032739] 2025-07-17T08:46:56.0329368Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:56.0332279Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_verifier.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:56.033057] 2025-07-17T08:46:59.1545355Z 2025-07-17T08:46:59.1546189Z export/test_verifier 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_verifier_1.1_8c8157ad505cce10_.log 2025-07-17T08:46:59.1546720Z Running 0 items in this shard: 2025-07-17T08:46:59.1546844Z 2025-07-17T08:46:59.1547011Z Running functorch/test_ac_knapsack 1/1 ... [2025-07-17 08:46:59.154528] 2025-07-17T08:46:59.1547295Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:46:59.1549684Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'functorch/test_ac_knapsack.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:46:59.154830] 2025-07-17T08:47:02.2766096Z 2025-07-17T08:47:02.2767073Z functorch/test_ac_knapsack 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ac_knapsack_1.1_7fda5c4d7dc49691_.log 2025-07-17T08:47:02.2767902Z Running 0 items in this shard: 2025-07-17T08:47:02.2768168Z 2025-07-17T08:47:02.2768416Z Running functorch/test_ac_logging 1/1 ... [2025-07-17 08:47:02.276545] 2025-07-17T08:47:02.2768891Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:47:02.2770312Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'functorch/test_ac_logging.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:47:02.276859] 2025-07-17T08:47:05.3485027Z 2025-07-17T08:47:05.3485734Z functorch/test_ac_logging 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ac_logging_1.1_bbf3472aea6b050a_.log 2025-07-17T08:47:05.3486242Z Running 0 items in this shard: 2025-07-17T08:47:05.3486381Z 2025-07-17T08:47:05.3486531Z Running functorch/test_control_flow 1/1 ... [2025-07-17 08:47:05.348525] 2025-07-17T08:47:05.3486809Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:47:05.3490037Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'functorch/test_control_flow.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:47:05.348834] 2025-07-17T08:47:09.3728253Z 2025-07-17T08:47:09.3728980Z functorch/test_control_flow 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_control_flow_1.1_7a25344a800898b4_.log 2025-07-17T08:47:09.3729507Z Running 0 items in this shard: 2025-07-17T08:47:09.3729633Z 2025-07-17T08:47:09.3729931Z Running inductor/test_aot_inductor_custom_ops 1/1 ... [2025-07-17 08:47:09.372818] 2025-07-17T08:47:09.3732729Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:47:09.3733464Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_aot_inductor_custom_ops.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:47:09.373147] 2025-07-17T08:47:17.0048957Z 2025-07-17T08:47:17.0057532Z inductor/test_aot_inductor_custom_ops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_custom_ops_1.1_c37b69de140aa8b3_.log 2025-07-17T08:47:17.0058113Z Running 0 items in this shard: 2025-07-17T08:47:17.0058238Z 2025-07-17T08:47:17.0058412Z Running inductor/test_aot_inductor_package 1/1 ... [2025-07-17 08:47:17.004901] 2025-07-17T08:47:17.0058703Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:47:17.0059393Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_aot_inductor_package.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:47:17.005218] 2025-07-17T08:47:24.3864439Z 2025-07-17T08:47:24.3865347Z inductor/test_aot_inductor_package 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_package_1.1_a0195c7a8b9b3dec_.log 2025-07-17T08:47:24.3865951Z Running 0 items in this shard: 2025-07-17T08:47:24.3866080Z 2025-07-17T08:47:24.3866246Z Running inductor/test_b2b_gemm 1/1 ... [2025-07-17 08:47:24.386448] 2025-07-17T08:47:24.3866519Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:47:24.3869596Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_b2b_gemm.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:47:24.386760] 2025-07-17T08:47:31.5434944Z 2025-07-17T08:47:31.5435946Z inductor/test_b2b_gemm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_b2b_gemm_1.1_59f2673917a325a7_.log 2025-07-17T08:47:31.5436624Z Running 0 items in this shard: 2025-07-17T08:47:31.5436849Z 2025-07-17T08:47:31.5437223Z Running inductor/test_compile_subprocess 2/2 ... [2025-07-17 08:47:31.543520] 2025-07-17T08:47:31.5437565Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:47:31.5439694Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_compile_subprocess.py', '-m', 'serial', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:47:31.543825] 2025-07-17T08:47:39.3758489Z 2025-07-17T08:47:39.3759329Z inductor/test_compile_subprocess 2/2 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_compile_subprocess_2.2_92d6d35945614169_.log 2025-07-17T08:47:39.3759895Z Running 0 items in this shard: 2025-07-17T08:47:39.3760028Z 2025-07-17T08:47:39.3760199Z Running inductor/test_compiled_optimizers 2/3 ... [2025-07-17 08:47:39.375765] 2025-07-17T08:47:39.3760496Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:47:39.3762347Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_compiled_optimizers.py', '-m', 'serial', '--shard-id=2', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:47:39.376077] 2025-07-17T08:47:48.0096214Z 2025-07-17T08:47:48.0097019Z inductor/test_compiled_optimizers 2/3 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_compiled_optimizers_2.3_5df44dde98880568_.log 2025-07-17T08:47:48.0097619Z Running 0 items in this shard: 2025-07-17T08:47:48.0097743Z 2025-07-17T08:47:48.0097940Z Running inductor/test_cuda_repro 1/1 ... [2025-07-17 08:47:48.009644] 2025-07-17T08:47:48.0098267Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:47:48.0100865Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_cuda_repro.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:47:48.009939] 2025-07-17T08:47:55.7917352Z 2025-07-17T08:47:55.7918861Z inductor/test_cuda_repro 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cuda_repro_1.1_f02cd4a99d7b5274_.log 2025-07-17T08:47:55.7919394Z Running 0 items in this shard: 2025-07-17T08:47:55.7919516Z 2025-07-17T08:47:55.7919674Z Running inductor/test_cudagraph_trees 1/1 ... [2025-07-17 08:47:55.791717] 2025-07-17T08:47:55.7919960Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:47:55.7921949Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_cudagraph_trees.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:47:55.792038] 2025-07-17T08:48:03.2226082Z 2025-07-17T08:48:03.2226985Z inductor/test_cudagraph_trees 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cudagraph_trees_1.1_688ac77279f760d9_.log 2025-07-17T08:48:03.2227561Z Running 0 items in this shard: 2025-07-17T08:48:03.2227689Z 2025-07-17T08:48:03.2228106Z Running inductor/test_cudagraph_trees_expandable_segments 1/1 ... [2025-07-17 08:48:03.222602] 2025-07-17T08:48:03.2228480Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:48:03.2230678Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_cudagraph_trees_expandable_segments.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:48:03.222910] 2025-07-17T08:48:10.5534879Z 2025-07-17T08:48:10.5535726Z inductor/test_cudagraph_trees_expandable_segments 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cudagraph_trees_expandable_segments_1.1_07d4f2d24e70ce91_.log 2025-07-17T08:48:10.5536384Z Running 0 items in this shard: 2025-07-17T08:48:10.5536511Z 2025-07-17T08:48:10.5536934Z Running inductor/test_custom_post_grad_passes 1/1 ... [2025-07-17 08:48:10.553504] 2025-07-17T08:48:10.5537285Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:48:10.5539956Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_custom_post_grad_passes.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:48:10.553804] 2025-07-17T08:48:17.9336211Z 2025-07-17T08:48:17.9337040Z inductor/test_custom_post_grad_passes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_custom_post_grad_passes_1.1_0fc2e864caa12fcf_.log 2025-07-17T08:48:17.9337606Z Running 0 items in this shard: 2025-07-17T08:48:17.9337733Z 2025-07-17T08:48:17.9337916Z Running inductor/test_cutlass_evt 1/1 ... [2025-07-17 08:48:17.933645] 2025-07-17T08:48:17.9338285Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:48:17.9341112Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_cutlass_evt.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:48:17.933948] 2025-07-17T08:48:25.3647805Z 2025-07-17T08:48:25.3648608Z inductor/test_cutlass_evt 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cutlass_evt_1.1_672b9b6faeed5fd2_.log 2025-07-17T08:48:25.3649127Z Running 0 items in this shard: 2025-07-17T08:48:25.3649259Z 2025-07-17T08:48:25.3649440Z Running inductor/test_debug_trace 1/1 ... [2025-07-17 08:48:25.364810] 2025-07-17T08:48:25.3652233Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:48:25.3653624Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_debug_trace.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:48:25.365109] 2025-07-17T08:48:33.0465376Z 2025-07-17T08:48:33.0466211Z inductor/test_debug_trace 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_debug_trace_1.1_c4767fd2656e8ff3_.log 2025-07-17T08:48:33.0466732Z Running 0 items in this shard: 2025-07-17T08:48:33.0466885Z 2025-07-17T08:48:33.0472686Z Running inductor/test_decompose_mem_bound_mm 1/1 ... [2025-07-17 08:48:33.046532] 2025-07-17T08:48:33.0473030Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:48:33.0473753Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_decompose_mem_bound_mm.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:48:33.046833] 2025-07-17T08:48:40.5285404Z 2025-07-17T08:48:40.5286231Z inductor/test_decompose_mem_bound_mm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_decompose_mem_bound_mm_1.1_0851e4a53b0f5a9f_.log 2025-07-17T08:48:40.5286930Z Running 0 items in this shard: 2025-07-17T08:48:40.5287056Z 2025-07-17T08:48:40.5287244Z Running inductor/test_dependencies 1/1 ... [2025-07-17 08:48:40.528598] 2025-07-17T08:48:40.5287526Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:48:40.5290485Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_dependencies.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:48:40.528892] 2025-07-17T08:48:47.8595160Z 2025-07-17T08:48:47.8595911Z inductor/test_dependencies 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_dependencies_1.1_4a3cfd090a920026_.log 2025-07-17T08:48:47.8596473Z Running 0 items in this shard: 2025-07-17T08:48:47.8596610Z 2025-07-17T08:48:47.8597273Z Running inductor/test_efficient_conv_bn_eval 1/1 ... [2025-07-17 08:48:47.859554] 2025-07-17T08:48:47.8597579Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:48:47.8600303Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_efficient_conv_bn_eval.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:48:47.859851] 2025-07-17T08:48:55.5415118Z 2025-07-17T08:48:55.5416360Z inductor/test_efficient_conv_bn_eval 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_efficient_conv_bn_eval_1.1_1acea78f0de28746_.log 2025-07-17T08:48:55.5417181Z Running 0 items in this shard: 2025-07-17T08:48:55.5417354Z 2025-07-17T08:48:55.5417599Z Running inductor/test_extension_backend 1/1 ... [2025-07-17 08:48:55.541504] 2025-07-17T08:48:55.5418048Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:48:55.5419608Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_extension_backend.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:48:55.541805] 2025-07-17T08:49:03.2735404Z 2025-07-17T08:49:03.2736129Z inductor/test_extension_backend 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_extension_backend_1.1_9f8a8fccf77ea7dd_.log 2025-07-17T08:49:03.2736684Z Running 0 items in this shard: 2025-07-17T08:49:03.2736815Z 2025-07-17T08:49:03.2737015Z Running inductor/test_external_callables 1/1 ... [2025-07-17 08:49:03.273556] 2025-07-17T08:49:03.2737322Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:49:03.2772648Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_external_callables.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:49:03.273856] 2025-07-17T08:49:06.8470293Z 2025-07-17T08:49:06.8471281Z inductor/test_external_callables 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_external_callables_1.1_c210444507bda5dc_.log 2025-07-17T08:49:06.8472087Z Running 0 items in this shard: 2025-07-17T08:49:06.8472261Z 2025-07-17T08:49:06.8473358Z Running inductor/test_flex_decoding 1/3 ... [2025-07-17 08:49:06.847187] 2025-07-17T08:49:06.8473768Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:49:06.8479279Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_flex_decoding.py', '-m', 'serial', '--shard-id=1', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:49:06.847532] 2025-07-17T08:49:10.8212862Z 2025-07-17T08:49:10.8213879Z inductor/test_flex_decoding 1/3 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_flex_decoding_1.3_711199c39cc17526_.log 2025-07-17T08:49:10.8214429Z Running 0 items in this shard: 2025-07-17T08:49:10.8214560Z 2025-07-17T08:49:10.8214694Z Running inductor/test_mps_basic 1/1 ... [2025-07-17 08:49:10.821259] 2025-07-17T08:49:10.8214984Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:49:10.8217529Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_mps_basic.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:49:10.821574] 2025-07-17T08:49:18.2524989Z 2025-07-17T08:49:18.2525650Z inductor/test_mps_basic 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_mps_basic_1.1_da6838f1ee92b667_.log 2025-07-17T08:49:18.2526110Z 2025-07-17T08:49:18.2526686Z Running inductor/test_ordered_set 1/1 ... [2025-07-17 08:49:18.252525] 2025-07-17T08:49:18.2526987Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:49:18.2531228Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_ordered_set.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:49:18.252827] 2025-07-17T08:49:21.4248140Z 2025-07-17T08:49:21.4248930Z inductor/test_ordered_set 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_ordered_set_1.1_8e34a76206f8a879_.log 2025-07-17T08:49:21.4249464Z Running 0 items in this shard: 2025-07-17T08:49:21.4249599Z 2025-07-17T08:49:21.4250197Z Running inductor/test_pad_mm 1/1 ... [2025-07-17 08:49:21.424866] 2025-07-17T08:49:21.4250492Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:49:21.4253626Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_pad_mm.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:49:21.425165] 2025-07-17T08:49:28.7057576Z 2025-07-17T08:49:28.7058331Z inductor/test_pad_mm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_pad_mm_1.1_7ccabad27ace694d_.log 2025-07-17T08:49:28.7058841Z Running 0 items in this shard: 2025-07-17T08:49:28.7058973Z 2025-07-17T08:49:28.7059321Z Running inductor/test_padding 1/1 ... [2025-07-17 08:49:28.705792] 2025-07-17T08:49:28.7059595Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:49:28.7064662Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_padding.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:49:28.706088] 2025-07-17T08:49:42.8524848Z 2025-07-17T08:49:42.8526228Z inductor/test_padding 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_padding_1.1_2c68ce3cfda7020e_.log 2025-07-17T08:49:42.8526934Z Running 1 items in this shard: test/inductor/test_padding.py::PaddingTest::test_nobias_LinearAndSoftmax_codegen 2025-07-17T08:49:42.8527238Z 2025-07-17T08:49:42.8527400Z Running inductor/test_pattern_matcher 1/1 ... [2025-07-17 08:49:42.852452] 2025-07-17T08:49:42.8527692Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:49:42.8528903Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_pattern_matcher.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:49:42.852752] 2025-07-17T08:49:50.0833699Z 2025-07-17T08:49:50.0834753Z inductor/test_pattern_matcher 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_pattern_matcher_1.1_d48fd310aac8d2fa_.log 2025-07-17T08:49:50.0835310Z Running 0 items in this shard: 2025-07-17T08:49:50.0835447Z 2025-07-17T08:49:50.0836202Z Running inductor/test_perf 1/1 ... [2025-07-17 08:49:50.083445] 2025-07-17T08:49:50.0841732Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:49:50.0842433Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_perf.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:49:50.083866] 2025-07-17T08:49:57.5153170Z 2025-07-17T08:49:57.5154038Z inductor/test_perf 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_perf_1.1_5ed4eea2809434c6_.log 2025-07-17T08:49:57.5154548Z Running 0 items in this shard: 2025-07-17T08:49:57.5154680Z 2025-07-17T08:49:57.5154826Z Running inductor/test_profiler 1/1 ... [2025-07-17 08:49:57.515310] 2025-07-17T08:49:57.5155105Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:49:57.5157977Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_profiler.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:49:57.515635] 2025-07-17T08:50:04.6458568Z 2025-07-17T08:50:04.6459822Z inductor/test_profiler 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_profiler_1.1_2fc46f39857bc075_.log 2025-07-17T08:50:04.6460353Z Running 0 items in this shard: 2025-07-17T08:50:04.6460486Z 2025-07-17T08:50:04.6460629Z Running inductor/test_quantization 1/1 ... [2025-07-17 08:50:04.645835] 2025-07-17T08:50:04.6460918Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:50:04.6462895Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_quantization.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:50:04.646153] 2025-07-17T08:50:12.0268176Z 2025-07-17T08:50:12.0269045Z inductor/test_quantization 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_quantization_1.1_91fd569d21204846_.log 2025-07-17T08:50:12.0269568Z Running 0 items in this shard: 2025-07-17T08:50:12.0269694Z 2025-07-17T08:50:12.0269879Z Running inductor/test_scatter_optimization 1/1 ... [2025-07-17 08:50:12.026749] 2025-07-17T08:50:12.0270185Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:50:12.0272603Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_scatter_optimization.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:50:12.027068] 2025-07-17T08:50:19.2064969Z 2025-07-17T08:50:19.2066422Z inductor/test_scatter_optimization 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_scatter_optimization_1.1_b75dea350acf8d05_.log 2025-07-17T08:50:19.2067009Z Running 0 items in this shard: 2025-07-17T08:50:19.2067140Z 2025-07-17T08:50:19.2067307Z Running inductor/test_select_algorithm 1/1 ... [2025-07-17 08:50:19.206463] 2025-07-17T08:50:19.2067610Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:50:19.2069441Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_select_algorithm.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:50:19.206761] 2025-07-17T08:50:26.5371475Z 2025-07-17T08:50:26.5372341Z inductor/test_select_algorithm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_select_algorithm_1.1_e23e38e5a5bdeb3f_.log 2025-07-17T08:50:26.5372904Z Running 0 items in this shard: 2025-07-17T08:50:26.5373032Z 2025-07-17T08:50:26.5373287Z Running inductor/test_snode_runtime 1/1 ... [2025-07-17 08:50:26.537146] 2025-07-17T08:50:26.5373589Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:50:26.5378781Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_snode_runtime.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:50:26.537470] 2025-07-17T08:50:33.7681027Z 2025-07-17T08:50:33.7681777Z inductor/test_snode_runtime 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_snode_runtime_1.1_063b9585341bcba3_.log 2025-07-17T08:50:33.7682325Z Running 0 items in this shard: 2025-07-17T08:50:33.7682449Z 2025-07-17T08:50:33.7684741Z Running inductor/test_split_cat_fx_aten_passes 1/1 ... [2025-07-17 08:50:33.768223] 2025-07-17T08:50:33.7685082Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:50:33.7688915Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_split_cat_fx_aten_passes.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:50:33.768638] 2025-07-17T08:50:41.1996678Z 2025-07-17T08:50:41.1997431Z inductor/test_split_cat_fx_aten_passes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_split_cat_fx_aten_passes_1.1_6d884d9c4ceafef7_.log 2025-07-17T08:50:41.2000794Z Running 0 items in this shard: 2025-07-17T08:50:41.2000964Z 2025-07-17T08:50:41.2001302Z Running inductor/test_split_cat_fx_passes 1/1 ... [2025-07-17 08:50:41.199682] 2025-07-17T08:50:41.2001787Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:50:41.2004049Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_split_cat_fx_passes.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:50:41.199986] 2025-07-17T08:50:48.5303380Z 2025-07-17T08:50:48.5304275Z inductor/test_split_cat_fx_passes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_split_cat_fx_passes_1.1_22d9869986bbb101_.log 2025-07-17T08:50:48.5304828Z Running 0 items in this shard: 2025-07-17T08:50:48.5304951Z 2025-07-17T08:50:48.5305135Z Running inductor/test_static_cuda_launcher 1/1 ... [2025-07-17 08:50:48.530281] 2025-07-17T08:50:48.5305441Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:50:48.5335132Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_static_cuda_launcher.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:50:48.530597] 2025-07-17T08:50:55.7111017Z 2025-07-17T08:50:55.7111742Z inductor/test_static_cuda_launcher 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_static_cuda_launcher_1.1_6e8fba75fc676c22_.log 2025-07-17T08:50:55.7112294Z Running 0 items in this shard: 2025-07-17T08:50:55.7112424Z 2025-07-17T08:50:55.7112585Z Running inductor/test_subgraph_choice 1/1 ... [2025-07-17 08:50:55.711082] 2025-07-17T08:50:55.7112871Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:50:55.7115376Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_subgraph_choice.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:50:55.711380] 2025-07-17T08:51:02.8414322Z 2025-07-17T08:51:02.8415394Z inductor/test_subgraph_choice 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_subgraph_choice_1.1_66b4e3bb2b31defc_.log 2025-07-17T08:51:02.8415949Z Running 0 items in this shard: 2025-07-17T08:51:02.8416079Z 2025-07-17T08:51:02.8416214Z Running inductor/test_torchbind 1/1 ... [2025-07-17 08:51:02.841406] 2025-07-17T08:51:02.8416492Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:51:02.8418427Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchbind.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:51:02.841705] 2025-07-17T08:51:10.0717548Z 2025-07-17T08:51:10.0718395Z inductor/test_torchbind 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchbind_1.1_ef5b31d5f3a783c5_.log 2025-07-17T08:51:10.0719019Z Running 0 items in this shard: 2025-07-17T08:51:10.0719146Z 2025-07-17T08:51:10.0719389Z Running inductor/test_torchinductor_codegen_config_overrides 1/1 ... [2025-07-17 08:51:10.071718] 2025-07-17T08:51:10.0719733Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:51:10.0721684Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor_codegen_config_overrides.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:51:10.072019] 2025-07-17T08:51:17.3019878Z 2025-07-17T08:51:17.3020927Z inductor/test_torchinductor_codegen_config_overrides 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_codegen_config_overrides_1.1_5ba0cb3b09cff0b3_.log 2025-07-17T08:51:17.3021651Z Running 0 items in this shard: 2025-07-17T08:51:17.3021865Z 2025-07-17T08:51:17.3022127Z Running inductor/test_torchinductor_opinfo 1/13 ... [2025-07-17 08:51:17.302012] 2025-07-17T08:51:17.3022527Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:51:17.3024560Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'serial', '--shard-id=1', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:51:17.302296] 2025-07-17T08:51:26.4371766Z 2025-07-17T08:51:26.4372770Z inductor/test_torchinductor_opinfo 1/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_1.13_d8830e8271d94b22_.log 2025-07-17T08:51:26.4373524Z Running 0 items in this shard: 2025-07-17T08:51:26.4373687Z 2025-07-17T08:51:26.4373957Z Running inductor/test_torchinductor_opinfo 8/13 ... [2025-07-17 08:51:26.437130] 2025-07-17T08:51:26.4374346Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:51:26.4378357Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'serial', '--shard-id=8', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:51:26.437444] 2025-07-17T08:51:35.4722358Z 2025-07-17T08:51:35.4723377Z inductor/test_torchinductor_opinfo 8/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_8.13_5b60d271fa261836_.log 2025-07-17T08:51:35.4723940Z Running 0 items in this shard: 2025-07-17T08:51:35.4724073Z 2025-07-17T08:51:35.4725920Z Running inductor/test_torchinductor_opinfo 11/13 ... [2025-07-17 08:51:35.472446] 2025-07-17T08:51:35.4726241Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:51:35.4732538Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'serial', '--shard-id=11', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:51:35.472933] 2025-07-17T08:51:44.5569075Z 2025-07-17T08:51:44.5570907Z inductor/test_torchinductor_opinfo 11/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_11.13_a32c92a2ebf8aeae_.log 2025-07-17T08:51:44.5571505Z Running 0 items in this shard: 2025-07-17T08:51:44.5571642Z 2025-07-17T08:51:44.5571771Z Running optim/test_swa_utils 1/1 ... [2025-07-17 08:51:44.556862] 2025-07-17T08:51:44.5572039Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:51:44.5573100Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'optim/test_swa_utils.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:51:44.557168] 2025-07-17T08:51:47.5778830Z 2025-07-17T08:51:47.5779960Z optim/test_swa_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/optim.test_swa_utils_1.1_dbb6f612b7eb00e8_.log 2025-07-17T08:51:47.5780412Z 2025-07-17T08:51:47.5780616Z Running test_out_dtype_op 1/1 ... [2025-07-17 08:51:47.577861] 2025-07-17T08:51:47.5781011Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:51:47.5785727Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_out_dtype_op.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:51:47.578160] 2025-07-17T08:51:51.3511648Z 2025-07-17T08:51:51.3512434Z test_out_dtype_op 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_out_dtype_op_1.1_a4e0f21222282feb_.log 2025-07-17T08:51:51.3512906Z Running 0 items in this shard: 2025-07-17T08:51:51.3513032Z 2025-07-17T08:51:51.3513178Z Running test_package 1/1 ... [2025-07-17 08:51:51.351140] 2025-07-17T08:51:51.3515670Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:51:51.3516331Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_package.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:51:51.351455] 2025-07-17T08:51:54.6231699Z 2025-07-17T08:51:54.6232336Z test_package 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_package_1.1_68331b7a5d58c05e_.log 2025-07-17T08:51:54.6232782Z Running 0 items in this shard: 2025-07-17T08:51:54.6232914Z 2025-07-17T08:51:54.6233058Z Running test_per_overload_api 1/1 ... [2025-07-17 08:51:54.623148] 2025-07-17T08:51:54.6233325Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:51:54.6236184Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_per_overload_api.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:51:54.623463] 2025-07-17T08:51:57.7450398Z 2025-07-17T08:51:57.7451071Z test_per_overload_api 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_per_overload_api_1.1_f81e020e894e94f8_.log 2025-07-17T08:51:57.7451556Z Running 0 items in this shard: 2025-07-17T08:51:57.7451713Z 2025-07-17T08:51:57.7451815Z Running test_prims 1/1 ... [2025-07-17 08:51:57.744985] 2025-07-17T08:51:57.7452058Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:51:57.7456937Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_prims.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:51:57.745308] 2025-07-17T08:52:02.3704525Z 2025-07-17T08:52:02.3705207Z test_prims 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_prims_1.1_7329abccd3578f5b_.log 2025-07-17T08:52:02.3705690Z Running 0 items in this shard: 2025-07-17T08:52:02.3705834Z 2025-07-17T08:52:02.3706317Z Running test_proxy_tensor 1/1 ... [2025-07-17 08:52:02.370448] 2025-07-17T08:52:02.3706570Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:52:02.3711130Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_proxy_tensor.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:52:02.370754] 2025-07-17T08:52:07.0457885Z 2025-07-17T08:52:07.0458655Z test_proxy_tensor 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_proxy_tensor_1.1_60257f7466bfbdcf_.log 2025-07-17T08:52:07.0459136Z Running 0 items in this shard: 2025-07-17T08:52:07.0459269Z 2025-07-17T08:52:07.0459406Z Running test_pruning_op 1/1 ... [2025-07-17 08:52:07.045748] 2025-07-17T08:52:07.0459702Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:52:07.0462210Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_pruning_op.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:52:07.046070] 2025-07-17T08:52:10.2171232Z 2025-07-17T08:52:10.2171973Z test_pruning_op 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_pruning_op_1.1_80f37fd5fbd1da7f_.log 2025-07-17T08:52:10.2172440Z Running 0 items in this shard: 2025-07-17T08:52:10.2172568Z 2025-07-17T08:52:10.2172764Z Running test_public_bindings 1/1 ... [2025-07-17 08:52:10.217078] 2025-07-17T08:52:10.2173079Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:52:10.2177487Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_public_bindings.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:52:10.217374] 2025-07-17T08:52:13.2888959Z 2025-07-17T08:52:13.2889672Z test_public_bindings 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_public_bindings_1.1_9c14fdd65abe833c_.log 2025-07-17T08:52:13.2890159Z Running 0 items in this shard: 2025-07-17T08:52:13.2890284Z 2025-07-17T08:52:13.2890739Z Running test_pytree 1/1 ... [2025-07-17 08:52:13.288882] 2025-07-17T08:52:13.2890984Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:52:13.2893758Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_pytree.py', '-m', 'serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:52:13.289185] 2025-07-17T08:52:16.5611130Z 2025-07-17T08:52:16.5611897Z test_pytree 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_pytree_1.1_22198c39f9c75417_.log 2025-07-17T08:52:16.5612963Z Running 0 items in this shard: 2025-07-17T08:52:16.5613093Z 2025-07-17T08:52:16.5613466Z Running test_quantization 1/3 ... [2025-07-17 08:52:16.561062] 2025-07-17T08:52:16.5613729Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:52:16.5615228Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_quantization.py', '-m', 'serial', '--shard-id=1', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:52:16.561367] 2025-07-17T08:52:21.7380932Z 2025-07-17T08:52:21.7381646Z test_quantization 1/3 was successful, full logs can be found in artifacts with path test/test-reports/test_quantization_1.3_872d8a7839f24642_.log 2025-07-17T08:52:21.7382116Z Running 0 items in this shard: 2025-07-17T08:52:21.7382246Z 2025-07-17T08:52:21.7382391Z Running test_quantization 3/3 ... [2025-07-17 08:52:21.738050] 2025-07-17T08:52:21.7382691Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:52:21.7385239Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_quantization.py', '-m', 'serial', '--shard-id=3', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:52:21.738347] 2025-07-17T08:52:26.6142396Z 2025-07-17T08:52:26.6143200Z test_quantization 3/3 was successful, full logs can be found in artifacts with path test/test-reports/test_quantization_3.3_1e13acdb8277dd99_.log 2025-07-17T08:52:26.6143682Z Running 0 items in this shard: 2025-07-17T08:52:26.6143812Z 2025-07-17T08:52:29.2358061Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T08:52:29.2359008Z import pkg_resources 2025-07-17T08:52:29.2648244Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T08:52:29.2649131Z import pkg_resources 2025-07-17T08:52:29.6276821Z Running dynamo/test_inline_and_install 1/1 ... [2025-07-17 08:52:29.627192] 2025-07-17T08:52:29.6277176Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:52:29.6278315Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_inline_and_install.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:52:29.627612] 2025-07-17T08:52:29.6475612Z Running functorch/test_ops 3/5 ... [2025-07-17 08:52:29.647228] 2025-07-17T08:52:29.6475948Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:52:29.6478462Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'functorch/test_ops.py', '-m', 'not serial', '--shard-id=3', '--num-shards=5', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:52:29.647586] 2025-07-17T08:54:09.1028906Z 2025-07-17T08:54:09.1032347Z dynamo/test_inline_and_install 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_inline_and_install_1.1_fc71809b013c2abe_.log 2025-07-17T08:54:09.1087524Z Running 184 items in this shard: test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_access_class_method_from_user_class_attr_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_access_class_method_from_user_class_builtin_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_byte_tensor_does_not_crash_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_capture_symbolic_tracing_simple_within_fake_mode_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_capture_symbolic_tracing_within_fake_mode_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_cond_free_variables_overlapping_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_cond_op_param_buffer_lifted_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_cond_raise_user_error_on_branch_args_mismatch_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_cond_raise_user_error_on_branch_return_multiple_tensors_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_cond_raise_user_error_on_branch_return_non_tensor_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_cond_raise_user_error_on_mismatch_return_length_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_cond_raise_user_error_on_mismatch_return_tensor_meta_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_cond_raise_user_error_on_missing_args_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_cond_raise_user_error_on_non_list_operands_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_cond_raise_user_error_on_non_tensor_operands_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_cond_raise_user_error_on_unsupported_pred_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_cond_supported_pred_types_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_constraint_violation_error_messages_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dataclass_input_output_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dict_return_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dict_return_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dupes_2_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dupes_2_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dupes_and_bypass_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dupes_and_bypass_reorder_with_non_tensor_arg_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dupes_and_bypass_reorder_with_non_tensor_arg_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dupes_and_bypass_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dupes_and_bypass_with_non_tensor_arg_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dupes_and_bypass_with_non_tensor_arg_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dupes_and_bypass_with_non_tensor_output_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dupes_and_bypass_with_non_tensor_output_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dupes_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dupes_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dynamic_slicing_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dynamic_slicing_invalid_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dynamic_slicing_simple_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dynamo_enum_in_tuple_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_dynamo_list_index_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_empty_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_enforce_equalities_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_compare_optimize_with_make_fx_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_cond_in_aten_symbolic_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_control_flow_with_getattr_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_decomp_asserts_bad_args_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_decomp_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_defaults_ok_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_dynamic_control_flow_error_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_dynamic_dim_cleanup_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_dynamic_dim_not_1_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_dynamic_dim_range_constraint_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_graph_bypass_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_graph_bypass_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_graph_with_complex_reorder_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_graph_with_complex_reorder_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_graph_with_list_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_graph_with_list_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_identity_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_masking_with_no_grad_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_meta_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_meta_val_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_mismatched_out_2_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_mismatched_out_2_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_mismatched_out_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_mismatched_out_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_module_specify_constraints_signature_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_multi_dynamic_dim_constraint_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_multi_dynamic_dim_unsafe_relationship_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_nn_module_stack_patched_module_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_no_raise_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_no_tensor_computation_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_pass_arg_by_name_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_pass_arg_by_name_star_args_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_persist_assert_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_preserve_constraints_as_metadata_tensor_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_preserves_nn_module_stack_for_get_attr_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_raise_guard_full_constraint_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_raise_guard_partial_constraint_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_raise_on_relationship_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_shape_control_flow_1_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_specialized_int_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_symbolic_shape_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_args_and_empty_kwargs_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_args_with_default_None_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_args_with_default_float_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_args_with_default_tensor_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_args_with_default_tuple_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_builtin_op_on_assume_constant_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_cond_branches_calling_methods_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_cond_closure_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_cond_dynamic_shape_pred_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_cond_with_closed_function_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_dict_values_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_free_function_and_class_method_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_free_function_and_class_method_multiarg_diff_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_free_function_and_class_method_multiarg_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_free_function_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_global_function_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_in_unspecialized_nn_module_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_list_nonzero_free_function_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_list_nonzero_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_method_on_module_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_method_on_module_invoke_twice_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_none_control_flow_free_func_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_none_control_flow_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_not_none_control_flow_free_func_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_not_none_control_flow_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_not_none_control_flow_pos_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_not_return_const_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_constant_tuple_nonzero_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_functools_wrapped_fn_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_functools_wrapped_method_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_kwargs_and_empty_args_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_kwargs_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_kwargs_with_default_None_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_kwargs_with_default_float_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_kwargs_with_default_tensor_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_kwargs_with_default_tuple_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_map_cond_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_map_zero_sized_tensor_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_map_zero_sized_tensor_suppress_errors_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_module_layer_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_nonzero_static_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_shallow_list_copy_with_side_effects_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_shallow_list_copy_wo_side_effects_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_stack_trace_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_symbool_inputs_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_export_with_wrapped_fn_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_exported_graph_serialization_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_func_return_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_func_return_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_fx_pytree_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_immutable_list_dict_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_input_container_type_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_invalid_input_global_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_invalid_input_global_multiple_access_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_invalid_input_nonlocal_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_invalid_input_unused_nonlocal_ok_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_list_contains_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_list_not_contains_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_list_unpack_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_list_unpack_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_map_cond_param_buffer_lifted_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_mixed_real_and_fake_inputs_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_multiple_outputs_op_with_evaluator_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_nested_cond_op_param_buffer_lifted_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_no_tensor_computation_2_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_no_tensor_computation_2_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_no_tensor_computation_fail_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_no_tensor_computation_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_not_functionalize_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_param_buffer_safe_from_mutation_recurse_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_param_buffer_safe_from_mutation_simple_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_pre_dispatch_simple_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_predispatch_with_for_out_dtype_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_predispatch_with_for_out_dtype_nested_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_predispatch_with_higher_order_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_predispatch_with_higher_order_nested_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_preserve_fx_node_metadata_graph_break_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_preserve_fx_node_metadata_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_preserve_fx_node_metadata_inline_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_preserve_fx_node_metadata_recompile_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_remove_redundant_dynamic_dim_in_error_message_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_retracibility_dict_container_inp_out_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_retracibility_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_retracibility_nested_list_out_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_round_dynamic_shapes_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_strict_fake_tensor_prop_real_tensors_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_subclass_parameters_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_sum_param_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_sym_contains_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_symbolic_tracing_within_fake_mode_with_constraints_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_symbolic_tracing_within_fake_mode_with_constraints_with_parameters_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_symbool_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_torch_inference_mode_ctx_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_trivial_constraint_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_uncaptured_higher_order_op_error_not_suppresed_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_untracked_inputs_in_constraints_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_zeroes_in_and_out_different_shape_on_test_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_zeroes_in_and_out_different_shape_on_test_with_aten_graph_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_zeroes_in_new_shape_scalar_out_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_zeroes_in_new_shape_scalar_out_permute_dupe_and_bypass_inline_and_install, test/dynamo/test_inline_and_install.py::InlineAndInstallExportTests::test_zeroes_in_new_shape_scalar_out_permute_inline_and_install 2025-07-17T08:54:09.1141383Z 2025-07-17T08:54:09.1141593Z Running inductor/test_torchinductor_codegen_dynamic_shapes 1/5 ... [2025-07-17 08:54:09.103241] 2025-07-17T08:54:09.1141936Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T08:54:09.1142663Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor_codegen_dynamic_shapes.py', '-m', 'not serial', '--shard-id=1', '--num-shards=5', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 08:54:09.103557] 2025-07-17T09:02:31.4553216Z 2025-07-17T09:02:31.4557027Z inductor/test_torchinductor_codegen_dynamic_shapes 1/5 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_codegen_dynamic_shapes_1.5_2d1a17f62a204e02_.log 2025-07-17T09:02:31.4664987Z Running 330 items in this shard: test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_adaptive_avg_pool2d1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_adaptive_avg_pool2d_low_prec_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_adaptive_avg_pool_errors_with_long_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_adaptive_avg_pool_with_output_size_0_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_angle_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_argmax_argmin_with_duplicates_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_as_strided_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_assert_size_stride_op_name_pass_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_avg_pool2d3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_avg_pool3d_backward2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_avg_pool3d_backward3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_avg_pool3d_backward4_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_avg_pool3d_backward_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_avg_pool_errors_with_uint_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bmm1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_int_int32_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_int_uint8_int8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_bucketize_nd_tiling_False_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_builtins_round_float_ndigits_neg_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_builtins_round_int_ndigits_pos_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_builtins_round_int_ndigits_zero_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_cat_unbacked_legacy_empty_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_cat_upcasting_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_chunk_recompiles_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_clamp_type_promotion_non_tensor_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_concat_add_inplace_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_config_option_dont_assume_alignment_cudagraphs_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_conv_backward_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_convolution1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_cumsum_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_custom_op_2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dist_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_div2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_div7_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_div9_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_div_zero_dim_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dropout2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dropout_deterministic_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dropout_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtype_mismatch_issue_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_bfloat16_bfloat16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_bfloat16_int16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_bfloat16_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_float16_float16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_float16_float32_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_float16_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_float32_float32_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_float64_float16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_float64_int8_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int16_int16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int32_bfloat16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int32_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_dtypeview_int64_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_embedding_bag_byte_unpack_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_empty1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_empty_strided_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_erfinv_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_exp2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_expm1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_fft_real_input_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_fill1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_fill2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_flip_cat_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_float_index_expression_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_fractional_max_pool2d2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_full_like_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_full_truncation_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_gather2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_generated_code_has_alignment_assert_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_graph_partition_pad_dynamic_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_horizonal_fusion1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_index3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_index_propagation_remainder_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_index_put_as_masked_fill_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_index_put_deterministic_fallback_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_index_tensor_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_inner_fn_str_and_stride_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_issue102546_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_large_offset_pointwise_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_like_channels_last_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_log1p_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_log_fp64_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_low_memory_max_pool_dilation_2_dim_2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_masked_scatter_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_max_pool2d2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_mean_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_mixed_mm2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_mixed_mm3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_multilayer_sum_low_prec_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_mutable_custom_op_fixed_layout2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_mutations_loop_fusion_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_new_ones_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_no_mega_fusion_during_lowering_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_bessel_j0_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_expm1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_hermite_polynomial_he_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_i1e_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_pointwise_xlogy_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_prod_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_randint_kernel_count_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_randn_with_dtype_and_device_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_reduction2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_reduction3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_reinterpret_dtypeview_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_remove_noop_copy_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_require_stride_expanded_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_reuse_buffers_with_aliasing_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_scalar_cpu_tensor_arg_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_scatter4_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_scatter5_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_scatter_add2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_scatter_reduce3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_sdpa_prefer_nd_tiling_False_use_block_ptr_False_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_sdpa_prefer_nd_tiling_False_use_block_ptr_True_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_sdpa_unaligned_mask_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_sgn_extremal_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_shape_prop_torch_ones_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_should_pad_bench_for_bmm_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_slice_mutation2_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_slice_view_with_graph_break_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_softmax_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_sort_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_sort_transpose_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_special_polygamma_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_split_cumsum_low_prec_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_split_reduction_dynamic_shape_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_split_with_list_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_squeeze1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_stride_preservation_with_stride_modifying_fx_pass_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_sum1_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_sum_int_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_tan_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_tanh_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_tensor_index_slice_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_tmp_not_defined_issue1_use_block_ptr_True_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_to_dtype_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_unbacked_floordiv_simplify_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_unbacked_floordiv_simplify_errors_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_unspec_inputs_int16_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_upsample_nearest2d_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_vectorized_ops_masked_var_novec_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_views3_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenCpuTests::test_zero_dim_reductions_dynamic_shapes_cpu, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_AllenaiLongformerBase_repro_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test__dyn_quant_matmul_4bit_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test__unsafe_masked_index_put_accumulate_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_adaptive_avg_pool_with_output_size_0_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_adaptive_max_pool2d2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_adaptive_max_pool2d3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_add_const_int_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_add_inplace_permuted_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_alexnet_prefix_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_any_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_aoti_eager_dtype_device_layout_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_aoti_eager_support_out_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_aoti_eager_support_str_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_arange1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_arange2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_arange4_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_arange5_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_argmax_argmin_with_duplicates_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_argmax_min_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_argmax_to_float_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_assert_alignment_op_name_fail_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_avg_pool2d4_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_avg_pool2d6_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_avg_pool2d_backward_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_avg_pool_errors_with_uint_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_batch_norm_2d_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_add_autotune_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_int16_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_int16_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_int16_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_int32_int16_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_int64_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_int8_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_int_int8_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_bucketize_nd_tiling_True_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_buffer_copied_in_graph_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_buffer_use_after_remove_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_builtins_round_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_builtins_round_int_ndigits_zero_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_cat_inplace_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_chunk_recompiles_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_clone_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_complex_memory_overlap_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_config_option_dont_assume_alignment_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_consecutive_split_cumprod_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_const_int32_to_float_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_constant_pad_2d_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_constant_pad_float64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_constant_pad_nd_inplace_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_conv3d_channels_last_use_block_ptr_True_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_cos_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_cummin_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_cumsum_pattern_matcher_issue_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_device_assert_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_diagonal_copy_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_div2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dont_constant_fold_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dropout3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dropout_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtype_sympy_expr_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_bfloat16_float16_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_bfloat16_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_bfloat16_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float32_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float32_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float64_float32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float64_float64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float64_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_float64_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int16_float16_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int16_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int16_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int32_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int64_bfloat16_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int64_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int64_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int8_float32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int8_int16_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_int8_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_uint8_bfloat16_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_dtypeview_uint8_float32_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_erfc_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_expm1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fallback_mutable_op_with_return_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fft_real_input_real_output_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fill1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_float16_to_int16_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fractional_max_pool2d2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fractional_max_pool2d3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fractional_max_pool2d5_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_functionalize_rng_wrappers_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_fuse_tiled_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_generated_code_has_alignment_assert_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_getitem_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_glu_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_graph_partition_no_inputs_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_graph_partition_refcount_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_graph_partition_scalar_inputs_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_horizonal_fusion1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_index2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_index_propagation_abs_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_index_put_reinplace_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_index_remainder_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_index_tensor_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_indirect_load_broadcast_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_inductor_assert_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_inductor_layout_optimization_input_mutations_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_inplace_mixed_dtype_ops_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_input_mutation2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_invalid_operand_issue1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_issue102546_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_large_pointwise_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_linspace2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_log2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_log_softmax_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_matmul_layer_norm_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_max_pool2d1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_max_pool2d_with_indices_backward3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_multilayer_prime_size_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_multilayer_sum_low_prec_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_nan_to_num_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_needs_contiguous_strides_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_neg_max_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_new_ones_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_no_op_reduction_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_nonzero_unbacked_refinement_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pad_cast_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_bessel_j0_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_chebyshev_polynomial_w_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_erfc_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_exp2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_hermite_polynomial_he_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_legendre_polynomial_p_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_log1p_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_log_ndtr_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_modified_bessel_i0_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_ndtr_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_polygamma_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_shifted_chebyshev_polynomial_u_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pointwise_spherical_bessel_j0_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_polar_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pow1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_pow_symfloat_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_prod_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_reduction5_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_remove_noop_copy_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_remove_noop_view_default_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_remove_noop_view_dtype_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_repeat_as_strided_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_replication_pad_errors_with_bool_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_scatter_add1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_scatter_add2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_scheduler_vertical_fusion1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sdpa_prefer_nd_tiling_False_use_block_ptr_False_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sdpa_prefer_nd_tiling_True_use_block_ptr_False_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_setitem_with_int_parameter_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_should_pad_bench_for_bmm_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sin_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_slice4_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_slice_mutation1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_slice_mutation2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_slice_scatter5_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_slice_scatter_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sort_bool_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_special_polygamma_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_split_cumprod_low_prec_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sqrt_dynamic_shapes_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_squeeze1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_squeeze2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_stride_preservation_with_stride_modifying_fx_pass_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_strided_inputs_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_sum_keepdims_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_tensor2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_tmp_not_defined_issue2_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_to_dtype_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_topk_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_unroll_small_reduction_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_unspec_inputs_int16_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_upsample_cat_conv_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_upsample_nearest3d_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_var_correction_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_var_mean_div_by_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_views1_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_views3_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_views5_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_where_broadcast_dynamic_shapes_cuda, test/inductor/test_torchinductor_codegen_dynamic_shapes.py::DynamicShapesCodegenGPUTests::test_xblock_divides_xnumel_dynamic_shapes_cuda 2025-07-17T09:02:31.4771704Z 2025-07-17T09:02:31.4771842Z Running test_nestedtensor 2/3 ... [2025-07-17 09:02:31.455956] 2025-07-17T09:02:31.4772108Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:02:31.4772756Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_nestedtensor.py', '-m', 'not serial', '--shard-id=2', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:02:31.456270] 2025-07-17T09:03:40.7050838Z 2025-07-17T09:03:40.7051560Z functorch/test_ops 3/5 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ops_3.5_6d0b8f5c3f5b53c4_.log 2025-07-17T09:03:40.7510392Z Running 2000 items in this shard: test/functorch/test_ops.py::TestOperatorsCUDA::test_data_write_errors_under_transform_cuda, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_SortGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad___rmul___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad__batch_norm_with_update_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad__native_batch_norm_legit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad__segment_reduce_offsets_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad__unsafe_masked_index_put_accumulate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_abs_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_acosh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_addbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_angle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_argwhere_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_as_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_atan2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_bool_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_broadcast_to_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_cartesian_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_cdist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_cfloat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_clamp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_clamp_max_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_column_stack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_complex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_cosh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_count_nonzero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_dist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_double_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_dstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_einsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_equal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_exp2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_expand_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_exponential_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_fft_fftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_fft_ifftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_fft_ihfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_fft_irfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_flatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_float_power_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_frexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_full_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_geometric_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_gt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_index_reduce_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_isclose_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_isinf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_isreal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_jiterator_4inputs_with_extra_args_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linalg_lstsq_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linalg_lu_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linalg_norm_subgradients_at_zero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linalg_qr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linalg_solve_triangular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_linalg_tensorsolve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_log10_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_log1p_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_logit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_logspace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_long_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_masked_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_masked_argmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_masked_log_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_masked_logaddexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_masked_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_masked_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_masked_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_masked_var_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_matrix_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_max_pool2d_with_indices_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_meshgrid_list_of_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_minimum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_movedim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_mvlgamma_mvlgamma_p_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_new_empty_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_new_ones_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nextafter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_adaptive_avg_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_adaptive_avg_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_conv1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_conv2d_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_conv2d_stride_depthwise_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_conv2d_stride_groups_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_dropout3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_dropout_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_hardsigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_hardswish_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_hardtanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_huber_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_max_unpool1d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_max_unpool3d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_pad_constant_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_pdist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_smooth_l1_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nn_functional_triplet_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_nonzero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_ones_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_pca_lowrank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_polygamma_polygamma_n_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_put_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_randn_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_ravel_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_remainder_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_repeat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_roll_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_round_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_scalar_tensor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_scatter_reduce_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_searchsorted_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_signal_windows_bartlett_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_signal_windows_exponential_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_signal_windows_hamming_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_signbit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_sort_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_bessel_j0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_bessel_y0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_bessel_y1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_chebyshev_polynomial_v_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_hermite_polynomial_h_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_hermite_polynomial_he_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_modified_bessel_i0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_ndtri_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_scaled_modified_bessel_k0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_special_zeta_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_squeeze_multiple_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_sum_to_size_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_tril_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_unbind_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_grad_unsqueeze_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_NumpyExpMarkDirtyAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_NumpyMulAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_NumpySortAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp___rmod___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp__softmax_backward_data_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp__unsafe_masked_index_put_accumulate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_acosh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_addbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_addcdiv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_addr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_argwhere_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_atanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_byte_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_ceil_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_conj_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_copysign_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_cos_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_cov_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_cummax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_cumsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_cumulative_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_deg2rad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_diff_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_div_floor_rounding_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_empty_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_empty_permuted_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_eq_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_erf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_expand_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_expand_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_expm1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_fft_irfft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_fft_rfft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_fft_rfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_flipud_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_floor_divide_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_gradient_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_half_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_index_add_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_index_put_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_index_reduce_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_jiterator_2inputs_2outputs_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_kron_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_le_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_lerp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_lgamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_cholesky_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_cholesky_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_eig_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_inv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_inv_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_ldl_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_lu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_matrix_rank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_pinv_hermitian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_svd_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_svdvals_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_linalg_tensorsolve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_log10_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_log1p_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_logcumsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_long_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_lu_unpack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_masked_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_masked_cumprod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_masked_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_masked_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_masked_std_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_matmul_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_meshgrid_list_of_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_minimum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_mode_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_movedim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_mv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_narrow_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_new_full_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_adaptive_avg_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_adaptive_max_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_alpha_dropout_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_batch_norm_without_cudnn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_conv1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_conv2d_stride_groups_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_conv_transpose2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_cosine_similarity_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_ctc_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_dropout2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_embedding_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_fractional_max_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_glu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_grid_sample_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_hardshrink_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_interpolate_area_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_interpolate_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_interpolate_trilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_leaky_relu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_linear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_local_response_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_max_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_multi_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_multilabel_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_pad_circular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_selu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_nn_functional_softmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_norm_nuc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_normal_number_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_ones_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_pinverse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_polygamma_polygamma_n_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_qr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_rand_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_randint_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_real_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_reshape_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_resize__cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_resize_as__cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_round_decimals_0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_round_decimals_neg_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_scatter_reduce_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_scatter_reduce_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_short_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_signal_windows_bartlett_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_signal_windows_blackman_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_sinh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_sparse_sampled_addmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_special_airy_ai_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_special_i0e_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_special_log_ndtr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_special_shifted_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_special_zeta_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_split_with_sizes_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_squeeze_multiple_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_std_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_take_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_tril_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_trunc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvp_unbind_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_H_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_NumpyTakeAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_ZeroGradientsGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp___rmatmul___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp___rmod___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp__chunk_cat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp__upsample_bilinear2d_aa_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_acos_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_acosh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_add_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_addbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_addmm_decomposed_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_argmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_as_strided_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_baddbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_bfloat16_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_bool_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_byte_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_chunk_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_conj_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_cos_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_digamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_dist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_eq_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_erfc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_erfinv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_expand_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_fft_fft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_fft_hfft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_fft_hfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_fft_ifftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_flip_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_flipud_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_float_power_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_fmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_full_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_index_add_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_index_reduce_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_int_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_isneginf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_jiterator_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_linalg_diagonal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_linalg_lu_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_linalg_lu_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_linalg_pinv_hermitian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_linalg_svd_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_linalg_vecdot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_logical_xor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_logspace_tensor_overload_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_long_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_mH_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_masked_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_masked_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_masked_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_masked_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_masked_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_masked_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_masked_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_matmul_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_max_reduction_no_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_maximum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_min_reduction_no_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_min_reduction_with_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_mode_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nan_to_num_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_native_batch_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_native_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_adaptive_avg_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_avg_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_batch_norm_without_cudnn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_celu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_conv2d_stride_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_interpolate_linear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_interpolate_nearest_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_logsigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_margin_ranking_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_selu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_unfold_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_upsample_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nn_functional_upsample_nearest_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_nonzero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_norm_inf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_normal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_ops_aten_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_polygamma_polygamma_n_0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_polygamma_polygamma_n_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_pow_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_rand_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_randint_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_randn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_ravel_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_resolve_conj_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_round_decimals_neg_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_scatter_reduce_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_short_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_sort_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_special_chebyshev_polynomial_w_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_special_hermite_polynomial_h_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_special_legendre_polynomial_p_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_special_scaled_modified_bessel_k0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_special_zeta_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_squeeze_multiple_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_stack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_std_mean_unbiased_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_sub_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_t_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_tan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_tensordot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_to_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_to_sparse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_trace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_tril_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_unflatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_unfold_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_uniform_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_unsafe_chunk_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_unsafe_split_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_unsqueeze_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_var_unbiased_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_vstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_where_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjp_zeros_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjpvmap_NumpyExpMarkDirtyAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjpvmap_NumpySortAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvjpvmap_SelectGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvmap_CubeGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvmap_NumpyTakeAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvmapvmap_NumpyCubeNotComposableAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvmapvmap_NumpyExpMarkDirtyAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvmapvmap_SelectAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_jvpvmapvmap_SortGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_diagonal_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_list_return_split_list_args_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_list_return_unbind_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_mH_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_mH_grad_op_vjp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_movedim_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_positive_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_real_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_resolve_conj_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_squeeze_multiple_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_unflatten_grad_op_vjp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_view_then_inplace_view_as_complex_grad_op_jvp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_ForwardHasDefaultArgsAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_NumpyTakeAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_SelectAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_ZeroGradientsGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp___rmul___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp___rsub___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp__batch_norm_with_update_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp__unsafe_masked_index_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp__upsample_bilinear2d_aa_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_addmv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_aminmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_argwhere_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_asinh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_atanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_atleast_2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_baddbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_bfloat16_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_broadcast_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_cartesian_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_cholesky_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_clamp_max_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_clamp_min_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_conj_physical_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_cumulative_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_div_trunc_rounding_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_double_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_dsplit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_expand_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_expm1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_fft_rfft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_flip_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_float_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_frac_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_full_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_geometric_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_geqrf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_index_reduce_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_inner_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_jiterator_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_eigh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_eigvals_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_eigvalsh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_inv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_inv_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_matrix_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_matrix_rank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_svd_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_vander_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_vecdot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_linalg_vector_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_long_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_lu_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_masked_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_masked_fill_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_masked_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_masked_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_masked_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_matrix_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_max_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_max_reduction_no_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_meshgrid_variadic_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_mvlgamma_mvlgamma_p_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_narrow_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_new_full_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_batch_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_conv2d_stride_padding_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_conv2d_strided_padding_dilation_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_grid_sample_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_interpolate_area_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_interpolate_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_interpolate_trilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_leaky_relu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_margin_ranking_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_max_unpool1d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_multi_head_attention_forward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_normalize_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_pad_replicate_negative_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_pairwise_distance_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_pixel_unshuffle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_relu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_nn_functional_softmin_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_norm_fro_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_norm_nuc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_normal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_normal_in_place_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_ones_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_ones_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_ops_aten_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_polygamma_polygamma_n_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_polygamma_polygamma_n_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_positive_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_rad2deg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_randint_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_randn_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_repeat_interleave_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_reshape_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_roll_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_scatter_reduce_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_scatter_reduce_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_signal_windows_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_signal_windows_gaussian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_signbit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_sparse_mm_reduce_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_special_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_special_hermite_polynomial_h_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_special_hermite_polynomial_he_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_split_list_args_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_sqrt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_stack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_std_unbiased_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_tile_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_to_sparse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_torch_ops_aten__efficient_attention_forward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_triangular_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_unbind_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_unflatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_unfold_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_unsafe_chunk_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_unsqueeze_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_var_unbiased_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_view_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjp_vstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_CubeGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_NumpyTakeAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_ZeroGradientsGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp___radd___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp___rmod___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp___rpow___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp__batch_norm_with_update_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp__native_batch_norm_legit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_abs_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_addbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_addcdiv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_addmm_decomposed_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_argsort_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_as_strided_partial_views_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_atan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_baddbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_bool_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_broadcast_shapes_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_broadcast_to_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_byte_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_cauchy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_cdouble_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_char_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_count_nonzero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_cumulative_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_empty_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_erfc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_expm1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_eye_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_fft_hfft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_float_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_float_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_gather_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_heaviside_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_hsplit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_hypot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_index_reduce_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_int_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_isnan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_isposinf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_cholesky_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_eig_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_eigvals_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_ldl_factor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_lstsq_grad_oriented_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_slogdet_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_solve_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_tensorsolve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_linalg_vector_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_log2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_log_softmax_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_logit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_long_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_lu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_mT_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_masked_cumprod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_masked_var_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_matmul_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_median_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nanquantile_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_native_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nextafter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_adaptive_avg_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_avg_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_conv2d_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_conv2d_stride_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_conv2d_strided_padding_dilation_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_dropout3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_fractional_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_instance_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_interpolate_bicubic_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_kl_div_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_linear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_max_unpool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_max_unpool1d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_max_unpool2d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_max_unpool3d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_normalize_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_pad_circular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_pixel_shuffle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_relu6_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_selu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_soft_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_softmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_softmin_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nn_functional_unfold_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_nonzero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_normal_number_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_ops_aten__new_zeros_with_same_feature_meta_functorchonly_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_put_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_rand_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_reciprocal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_round_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_sigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_signal_windows_bartlett_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_signal_windows_general_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_sinh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_sparse_sampled_addmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_bessel_y1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_chebyshev_polynomial_v_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_entr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_erfcx_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_i0e_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_legendre_polynomial_p_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_ndtr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_scaled_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_special_zeta_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_squeeze_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_squeeze_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_stack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_std_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_std_mean_unbiased_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_to_sparse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_torch_ops_aten__efficient_attention_forward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_trace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_unbind_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_unflatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_unsqueeze_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_zero__cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjp_zeros_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjpvmap_ForwardHasDefaultArgsAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjpvmap_NumpyCubeNotComposableAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvjpvmap_ZeroGradientsGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_NumpyCubeNotComposableAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_NumpyMulAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_NumpySortAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_SelectGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap___rpow___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_acos_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_addcdiv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_argsort_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_argwhere_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_asin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_asinh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_atan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_atleast_1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_bucketize_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_byte_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_chalf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_char_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_char_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_column_stack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_complex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_cos_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_cosh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_cov_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_cummin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_cumulative_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_diag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_diag_embed_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_digamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_double_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_empty_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_equal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_erfinv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_exp2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_expand_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_exponential_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_fft_fft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_fft_fftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_fft_fftshift_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_fft_hfftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_fft_irfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_floor_divide_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_fmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_frac_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_full_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_geometric_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_hstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_igamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_index_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_index_reduce_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_isfinite_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_jiterator_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_kron_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_ldexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_linalg_cholesky_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_linalg_cond_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_linalg_det_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_linalg_inv_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_linalg_ldl_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_linalg_lu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_linalg_lu_factor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_linalg_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_linalg_solve_triangular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_log_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_log_softmax_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_logcumsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_logical_or_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_long_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_lu_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_lu_unpack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_mH_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_masked_cumsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_masked_normalize_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_masked_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_masked_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_max_reduction_no_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_min_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_min_reduction_with_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_native_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_ne_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_neg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_avg_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_conv2d_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_conv2d_stride_depthwise_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_conv2d_stride_groups_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_conv3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_conv_transpose2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_dropout2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_gaussian_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_interpolate_trilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_leaky_relu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_local_response_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_logsigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_max_unpool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_mse_loss_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_multi_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_multilabel_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_pad_circular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_pad_reflect_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_pad_replicate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_pad_replicate_negative_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_pixel_shuffle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_triplet_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_nn_functional_upsample_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_norm_inf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_ops_aten_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_pca_lowrank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_pinverse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_polar_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_positive_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_rad2deg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_reciprocal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_remainder_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_renorm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_reshape_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_reshape_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_round_decimals_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_scatter_reduce_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_scatter_reduce_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_scatter_reduce_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_sigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_signal_windows_blackman_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_signal_windows_general_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_signal_windows_hamming_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_sin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_slice_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_softmax_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_bessel_y1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_log_ndtr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_modified_bessel_i0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_modified_bessel_i1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_special_shifted_chebyshev_polynomial_u_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_split_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_t_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_take_along_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_trace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_transpose_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_unbind_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_unbind_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_uniform_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_unsafe_split_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_var_unbiased_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmap_view_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmapvmap_CubeGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vjpvmapvmap_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_CubeGenVmapAutogradFunction_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_H_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_NumpyCubeNotComposableAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_NumpyExpMarkDirtyAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_ScaleGradGenVmapAutogradFunction_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_SortGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_T_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad___rdiv___cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad___rmatmul___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad___rmatmul___cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad___rmod___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad___rmod___cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad__batch_norm_with_update_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad__native_batch_norm_legit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad__segment_reduce_offsets_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad__softmax_backward_data_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad__softmax_backward_data_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad__unsafe_masked_index_put_accumulate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad__upsample_bilinear2d_aa_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_add_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_addmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_addmv_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_addr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_angle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_argmax_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_as_strided_copy_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_atleast_1d_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_broadcast_to_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_broadcast_to_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cauchy_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cdist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cdist_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cdouble_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cfloat_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_chalf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_char_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cholesky_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cholesky_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cholesky_solve_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_clamp_max_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_clone_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cosh_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_count_nonzero_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cummax_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cummin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_cumulative_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_deg2rad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_deg2rad_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_diff_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_div_floor_rounding_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_div_no_rounding_mode_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_dot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_dsplit_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_dstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_empty_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_empty_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_eq_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_equal_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_exp2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_expand_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_expm1_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_eye_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_eye_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_fft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_fftshift_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_fftshift_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_hfft_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_ifft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_ifft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_ifft_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_ifftshift_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_ihfft2_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_ihfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_ihfftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_rfft2_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_fft_rfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_flipud_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_float_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_float_power_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_floor_divide_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_full_like_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_grid_sampler_2d_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_gt_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_heaviside_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_histc_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_hstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_hypot_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_index_reduce_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_index_reduce_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_index_select_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_int_functorch_no_channels_last_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_isin_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_isnan_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_isreal_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_jiterator_2inputs_2outputs_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_jiterator_binary_return_by_ref_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_jiterator_unary_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_kthvalue_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_det_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_eigh_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_householder_product_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_inv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_inv_ex_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_ldl_factor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_ldl_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_lu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_matrix_rank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_matrix_rank_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_matrix_rank_hermitian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_norm_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_norm_subgradients_at_zero_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_pinv_singular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_qr_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_slogdet_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_solve_ex_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_solve_triangular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_svdvals_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_vecdot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linalg_vecdot_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_linspace_tensor_overload_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_log10_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_log2_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_log_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logaddexp2_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logical_and_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logical_and_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logical_or_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_logspace_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_lu_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_mH_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_mH_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_mT_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_amax_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_cumsum_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_log_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_median_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_scatter_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_masked_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_max_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_max_pool2d_with_indices_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_max_pool2d_with_indices_backward_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_max_reduction_with_dim_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_maximum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_meshgrid_list_of_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_mm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_mode_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_mode_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_mul_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_multinomial_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_mv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_mvlgamma_mvlgamma_p_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_mvlgamma_mvlgamma_p_5_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nanquantile_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nansum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_native_batch_norm_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_neg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_neg_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_new_empty_strided_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_adaptive_avg_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_alpha_dropout_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_avg_pool2d_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_avg_pool3d_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_batch_norm_without_cudnn_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_conv2d_no_bias_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_conv2d_strided_padding_dilation_no_bias_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_conv2d_strided_padding_dilation_with_bias_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_conv3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_conv_transpose2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_conv_transpose3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_conv_transpose3d_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_cosine_embedding_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_dropout3d_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_embedding_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_fractional_max_pool3d_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_gaussian_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_gelu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_gelu_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_glu_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_hardshrink_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_hardshrink_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_hardswish_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_hardtanh_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_instance_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_interpolate_area_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_interpolate_bicubic_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_interpolate_bilinear_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_interpolate_nearest-exact_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_margin_ranking_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_margin_ranking_loss_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_max_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_max_unpool1d_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_max_unpool2d_grad_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_mse_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_multi_head_attention_forward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_multi_margin_loss_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_multilabel_soft_margin_loss_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_normalize_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_pad_circular_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_pairwise_distance_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_pairwise_distance_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_pdist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_pixel_unshuffle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_poisson_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_poisson_nll_loss_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_rms_norm_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_selu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_softmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_softplus_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_softsign_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_threshold_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_triplet_margin_loss_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_unfold_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_nn_functional_upsample_nearest_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_norm_fro_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_norm_inf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_norm_nuc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_ones_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_ops_aten_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_outer_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_polygamma_polygamma_n_0_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_polygamma_polygamma_n_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_polygamma_polygamma_n_3_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_randint_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_randint_like_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_randn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_ravel_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_real_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_remainder_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_rot90_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_round_decimals_0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_scalar_tensor_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_scatter_add_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_scatter_reduce_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_scatter_reduce_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_scatter_reduce_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_scatter_reduce_prod_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_scatter_reduce_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_searchsorted_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_select_scatter_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_short_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_sigmoid_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_sign_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_signal_windows_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_signal_windows_exponential_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_signal_windows_general_hamming_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_signal_windows_nuttall_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_softmax_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_sparse_mm_reduce_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_airy_ai_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_bessel_j0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_chebyshev_polynomial_t_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_hermite_polynomial_h_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_i0e_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_laguerre_polynomial_l_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_legendre_polynomial_p_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_modified_bessel_i0_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_ndtri_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_shifted_chebyshev_polynomial_w_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_special_xlog1py_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_split_list_args_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_sqrt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_sqrt_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_squeeze_multiple_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_stack_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_std_unbiased_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_svd_lowrank_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_torch_ops_aten__safe_softmax_default_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_trace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_transpose_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_trapezoid_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_trapz_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_unflatten_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_uniform_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_unique_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_unsafe_split_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_unsafe_split_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_unsqueeze_copy_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_unsqueeze_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_view_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_vstack_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_where_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmap_autograd_grad_xlogy_cuda_float64, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_NumpyExpMarkDirtyAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall___radd___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall__chunk_cat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall__segment_reduce_offsets_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_acos_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_acosh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_addcmul_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_any_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_argmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_as_strided_partial_views_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_asin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_atan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_atleast_2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_atleast_3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_cfloat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_chalf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_char_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_clamp_min_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_clone_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_cov_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_cumsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_diag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_digamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_dist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_dot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_empty_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_erfc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_exp2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_expand_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_expand_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_fft_fft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_fft_fft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_fft_hfft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_fft_ifft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_fft_ifft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_fft_ihfftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_fft_irfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_full_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_CubeGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_MulGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_NumpyCubeAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule___getitem___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule___rsub___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule__native_batch_norm_legit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule__unsafe_masked_index_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_addbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_addr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_all_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_any_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_argmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_argwhere_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_asinh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_atleast_2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_atleast_3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_cat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_chalf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_clamp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_clamp_min_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_column_stack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_constant_pad_nd_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_cov_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_cummin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_cumsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_diagonal_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_diagonal_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_div_no_rounding_mode_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_dot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_double_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_empty_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_empty_permuted_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_eye_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_fft_hfftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_fft_ifft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_fft_ihfft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_float_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_fmod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_geometric_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_grid_sampler_2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_i0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_igamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_isreal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_item_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_ldexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_lerp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_cholesky_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_cond_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_inv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_ldl_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_lstsq_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_matrix_power_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_matrix_rank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_pinv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linalg_tensorsolve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_linspace_tensor_overload_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_log_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_logdet_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_logit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_logspace_tensor_overload_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_long_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_lu_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_masked_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_masked_cumsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_masked_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_masked_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_masked_normalize_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_masked_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_masked_softmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_masked_var_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_matrix_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_max_pool2d_with_indices_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_meshgrid_variadic_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_min_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_min_reduction_with_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_mode_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_mvlgamma_mvlgamma_p_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_native_dropout_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_new_full_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nextafter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_adaptive_max_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_avg_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_conv1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_conv2d_stride_padding_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_conv2d_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_conv_transpose1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_conv_transpose2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_dropout3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_dropout_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_elu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_embedding_bag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_embedding_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_hardswish_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_hardtanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_l1_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_linear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_max_unpool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_mish_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_multi_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_rms_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_normal_number_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_ops_aten_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_permute_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_positive_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_real_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_repeat_interleave_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_resolve_neg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_scatter_reduce_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_select_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_short_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_short_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_sigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_signal_windows_blackman_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_signal_windows_gaussian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_signal_windows_general_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_signal_windows_general_hamming_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_signal_windows_hann_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_sinh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_special_bessel_j1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_special_bessel_y0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_special_bessel_y1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_special_erfcx_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_special_hermite_polynomial_h_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_special_hermite_polynomial_he_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_special_legendre_polynomial_p_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_special_modified_bessel_k0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_square_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_sum_to_size_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_t_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_tensor_split_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_to_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_unbind_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_unique_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_unsafe_chunk_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_var_unbiased_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_has_batch_rule_where_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_hstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_index_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_index_reduce_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_index_reduce_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_index_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_jiterator_unary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_kthvalue_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_cond_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_inv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_inv_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_lstsq_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_lstsq_grad_oriented_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_matrix_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_matrix_rank_hermitian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_pinv_singular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_linalg_vander_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_log_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_logical_xor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_masked_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_masked_fill_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_masked_fill_functorch_Scalar_only_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_masked_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_masked_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_masked_std_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_max_reduction_no_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_max_reduction_with_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_maximum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_min_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_minimum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_msort_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_mv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_mvlgamma_mvlgamma_p_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nanmean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nanmedian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nanquantile_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nansum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_native_batch_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_native_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_conv1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_conv2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_conv2d_stride_groups_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_dropout_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_gaussian_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_hinge_embedding_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_huber_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_instance_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_interpolate_area_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_interpolate_nearest-exact_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_interpolate_nearest_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_leaky_relu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_linear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_local_response_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_margin_ranking_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_max_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_pad_circular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_pad_constant_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_pairwise_distance_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_relu6_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_rms_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_soft_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_nn_functional_triplet_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_permute_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_pinverse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_rad2deg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_randint_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_randint_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_randn_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_repeat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_resolve_neg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_rsqrt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_short_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_signal_windows_exponential_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_signal_windows_general_hamming_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_signal_windows_hamming_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_sinh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_slice_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_sparse_mm_reduce_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_chebyshev_polynomial_w_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_legendre_polynomial_p_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_scaled_modified_bessel_k0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_scaled_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_shifted_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_shifted_chebyshev_polynomial_w_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_spherical_bessel_j0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_xlog1py_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_special_zeta_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_squeeze_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_std_mean_unbiased_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_sub_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_svd_lowrank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_tanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_tile_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_triangular_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_unbind_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_unique_consecutive_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_view_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpall_zeros_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_CubeGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_NumpyCubeNotComposableAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_NumpySortAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_SortGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_T_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp___getitem___functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp___rdiv___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp___rmul___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp__segment_reduce_lengths_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp__softmax_backward_data_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_addbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_addcdiv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_alias_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_all_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_argwhere_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_as_strided_partial_views_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_asin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_bernoulli_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_block_diag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_byte_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_cdouble_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_char_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_clamp_min_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_constant_pad_nd_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_cov_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_cross_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_cumulative_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_diag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_diagonal_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_diff_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_div_no_rounding_mode_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_double_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_dstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_einsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_empty_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_expm1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_fft_fft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_fft_ifft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_fft_ihfftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_fft_irfftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_fft_rfftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_float_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_fmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_geometric_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_gt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_histc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_index_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_index_reduce_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_index_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_int_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_isclose_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_isfinite_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_isinf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_item_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_jiterator_unary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_kthvalue_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_cholesky_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_diagonal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_eigh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_eigvals_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_lu_factor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_linalg_pinv_hermitian_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_logaddexp2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_logdet_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_logical_not_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_logical_or_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_long_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_lu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_lu_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_lu_unpack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_mH_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_masked_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_masked_logaddexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_masked_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_max_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_mode_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_multinomial_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nanmean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_batch_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_celu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_channel_shuffle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_conv1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_conv2d_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_conv_transpose2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_cosine_embedding_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_dropout_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_fractional_max_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_grid_sample_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_instance_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_interpolate_bicubic_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_interpolate_trilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_max_unpool1d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_max_unpool3d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_mse_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_mse_loss_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_multi_head_attention_forward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_multilabel_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_pixel_unshuffle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_selu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_silu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_tanhshrink_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_nn_functional_triplet_margin_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_norm_inf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_normal_in_place_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_ops_aten__new_zeros_with_same_feature_meta_functorchonly_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_ops_aten_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_outer_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_pca_lowrank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_permute_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_polygamma_polygamma_n_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_polygamma_polygamma_n_4_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_positive_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_qr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_quantile_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_randn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_ravel_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_reshape_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_roll_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_scatter_reduce_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_short_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_sign_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_signal_windows_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_signal_windows_exponential_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_signal_windows_general_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_sinc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_sparse_sampled_addmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_bessel_j0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_bessel_y0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_chebyshev_polynomial_v_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_entr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_i1e_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_laguerre_polynomial_l_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_modified_bessel_i0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_modified_bessel_i1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_special_scaled_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_tanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_tensordot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_to_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_to_sparse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_topk_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_torch_ops_aten__efficient_attention_forward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_triangular_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_true_divide_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_unbind_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_unsafe_chunk_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_unsafe_split_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_var_unbiased_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvjp_view_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapjvpvmap_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_MulGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_NumpyExpMarkDirtyAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_NumpyTakeAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_ScaleGradGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_T_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp___rdiv___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp__chunk_cat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp__segment_reduce_offsets_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp__unsafe_masked_index_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp__unsafe_masked_index_put_accumulate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_addbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_addmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_argmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_as_strided_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_as_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_bernoulli_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_bool_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_byte_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_char_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_char_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_cholesky_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_clone_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_cov_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_cross_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_cummin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_cumulative_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_deg2rad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_diag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_diagonal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_diagonal_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_digamma_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_fft_fft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_fft_ifft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_fft_ifftn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_fft_ihfft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_fft_ihfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_fft_rfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_flatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_float_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_float_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_float_power_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_floor_divide_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_frac_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_frexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_geometric_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_gt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_NumpyCubeAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_NumpyMulAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_NumpyTakeAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_SortGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule___rdiv___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule___rmatmul___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule___rmul___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule__batch_norm_with_update_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule__unsafe_masked_index_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_angle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_arange_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_argsort_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_baddbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_block_diag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_broadcast_shapes_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_byte_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_cat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_char_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_cholesky_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_cholesky_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_copysign_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_corrcoef_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_diag_embed_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_div_floor_rounding_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_div_no_rounding_mode_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_empty_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_exp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_expm1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_fft_fft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_fft_ifftshift_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_float_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_fmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_full_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_gt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_half_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_half_functorch_no_channels_last_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_histc_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_index_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_index_reduce_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_isinf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_isnan_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_isposinf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_jiterator_4inputs_with_extra_args_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_kthvalue_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_le_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_cross_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_det_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_eigvals_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_lu_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_norm_subgradients_at_zero_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_pinv_singular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_solve_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_linalg_svdvals_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_log10_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_logdet_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_logical_or_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_logspace_tensor_overload_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_long_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_masked_amin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_masked_argmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_masked_fill_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_masked_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_masked_normalize_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_max_pool2d_with_indices_backward_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_max_reduction_no_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_maximum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_meshgrid_variadic_tensors_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_mvlgamma_mvlgamma_p_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_mvlgamma_mvlgamma_p_5_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nansum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_new_ones_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_batch_norm_without_cudnn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_conv2d_stride_depthwise_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_conv2d_stride_padding_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_conv2d_strided_padding_dilation_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_cosine_similarity_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_embedding_bag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_gaussian_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_gelu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_glu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_interpolate_area_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_kl_div_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_l1_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_linear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_max_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_max_unpool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_normalize_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_relu6_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_rms_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_smooth_l1_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_softmin_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_softplus_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_softshrink_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nn_functional_upsample_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_nonzero_static_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_permute_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_polygamma_polygamma_n_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_polygamma_polygamma_n_2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_pow_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_randint_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_randn_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_remainder_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_renorm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_roll_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_round_decimals_3_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_scalar_tensor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_scatter_add_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_sgn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_sigmoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_signal_windows_general_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_signal_windows_nuttall_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_bessel_j0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_bessel_j1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_hermite_polynomial_h_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_hermite_polynomial_he_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_legendre_polynomial_p_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_log_ndtr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_modified_bessel_i1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_scaled_modified_bessel_k0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_shifted_chebyshev_polynomial_u_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_special_zeta_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_split_with_sizes_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_std_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_svd_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_svd_lowrank_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_tanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_triangular_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_unbind_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_unflatten_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_unsafe_split_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_var_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_vsplit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_has_batch_rule_vstack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_igammac_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_index_reduce_prod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_int_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_kthvalue_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_le_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_det_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_eig_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_ldl_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_lstsq_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_lu_factor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_multi_dot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linalg_solve_triangular_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linspace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_linspace_tensor_overload_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_log10_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_log2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_log_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_logcumsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_logical_or_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_lu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_mT_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_masked_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_masked_cumprod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_masked_log_softmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_masked_logaddexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_masked_logsumexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_masked_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_masked_std_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_masked_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_max_reduction_no_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_max_reduction_with_dim_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_msort_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_mvlgamma_mvlgamma_p_1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nanmean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_narrow_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_native_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_ne_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_new_empty_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_adaptive_avg_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_avg_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_batch_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_conv2d_stride_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_conv2d_stride_padding_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_conv2d_stride_with_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_conv2d_strided_padding_dilation_no_bias_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_cosine_similarity_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_ctc_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_dropout_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_embedding_bag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_gaussian_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_glu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_hardshrink_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_hardswish_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_hardtanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_hinge_embedding_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_interpolate_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_interpolate_linear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_kl_div_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_margin_ranking_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_max_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_max_unpool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_mish_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_prelu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_relu6_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_nn_functional_silu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_normal_in_place_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_ones_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_ops_aten_index_put_functorch_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_pinverse_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_quantile_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_randint_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_randn_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_reshape_as_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_rot90_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_round_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_scalar_tensor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_scatter_reduce_amax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_scatter_reduce_sum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_select_scatter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_bessel_j1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_chebyshev_polynomial_v_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_i1e_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_log_ndtr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_ndtr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_ndtri_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_scaled_modified_bessel_k0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_scaled_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_shifted_chebyshev_polynomial_v_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_special_shifted_chebyshev_polynomial_w_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_squeeze_multiple_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_trace_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_triangular_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_unbind_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_unsafe_chunk_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_var_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_where_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjp_zeros_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_CubeGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_MulGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_SortGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp___getitem___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp___rsub___cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp__unsafe_masked_index_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp__unsafe_masked_index_put_accumulate_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_acosh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_addbmm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_addcdiv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_addmv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_aminmax_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_arange_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_argmin_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_argwhere_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_as_strided_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_as_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_asinh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_atleast_2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_bfloat16_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_byte_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_cat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_cdist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_cfloat_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_clamp_max_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_column_stack_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_complex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_cos_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_cumprod_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_cumsum_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_cumulative_trapezoid_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_diag_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_diagonal_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_empty_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_empty_strided_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_erfinv_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_expand_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_fft_fft2_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_fft_rfft_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_fliplr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_floor_divide_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_frac_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_frexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_gradient_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_igammac_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_index_fill_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_index_select_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_inner_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_isinf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_item_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_jiterator_unary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_kron_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_linalg_eig_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_linalg_ldl_solve_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_linalg_lu_factor_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_linalg_lu_factor_ex_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_linalg_matrix_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_linalg_multi_dot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_linalg_svdvals_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_linalg_vander_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_linalg_vecdot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_log_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_log_normal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_logaddexp_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_logical_or_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_lt_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_masked_fill_functorch_Scalar_only_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_masked_median_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_max_binary_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_multinomial_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_ne_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_neg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_new_empty_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nextafter_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_adaptive_avg_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_adaptive_max_pool1d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_avg_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_batch_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_bilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_binary_cross_entropy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_conv3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_dropout2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_elu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_glu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_grid_sample_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_hardshrink_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_hinge_embedding_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_interpolate_linear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_interpolate_trilinear_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_layer_norm_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_max_pool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_max_pool3d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_max_unpool1d_grad_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_max_unpool2d_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_mse_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_pad_reflect_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_pdist_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_pixel_unshuffle_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_poisson_nll_loss_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_rrelu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_nn_functional_softmin_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_norm_inf_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_normal_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_rad2deg_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_rand_like_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_resize_as__cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_roll_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_rot90_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_round_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_round_decimals_0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_scatter_reduce_mean_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_searchsorted_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_short_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_sign_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_signal_windows_cosine_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_signbit_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_softmax_with_dtype_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_airy_ai_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_bessel_y0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_chebyshev_polynomial_u_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_entr_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_modified_bessel_i0_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_modified_bessel_i1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_scaled_modified_bessel_k1_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_shifted_chebyshev_polynomial_t_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_shifted_chebyshev_polynomial_u_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_shifted_chebyshev_polynomial_v_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_special_zeta_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_split_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_split_list_args_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_split_with_sizes_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_sub_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_tanh_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_tensor_split_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_tensordot_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_topk_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_torch_ops_aten__safe_softmax_default_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_transpose_copy_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_triu_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_unfold_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_var_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_view_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvjp_where_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvmap_ForwardHasDefaultArgsAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvmap_MulGenVmapAutogradFunction_cuda_float32, test/functorch/test_ops.py::TestOperatorsCUDA::test_vmapvjpvmap_NumpyCubeAutogradFunction_cuda_float32 2025-07-17T09:03:40.7940528Z 2025-07-17T09:03:40.7940747Z Running inductor/test_torchinductor_dynamic_shapes 2/6 ... [2025-07-17 09:03:40.707512] 2025-07-17T09:03:40.7941396Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:03:40.7942119Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor_dynamic_shapes.py', '-m', 'not serial', '--shard-id=2', '--num-shards=6', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:03:40.707870] 2025-07-17T09:09:36.1685092Z 2025-07-17T09:09:36.1689449Z test_nestedtensor 2/3 was successful, full logs can be found in artifacts with path test/test-reports/test_nestedtensor_2.3_18ad6bea000919eb_.log 2025-07-17T09:09:36.1858195Z Running 525 items in this shard: test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_2_max_seq_len_3_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_2d_nested_tensor_batch_size_4_max_seq_len_5_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_batch_size_4_max_seq_len_3_vocab_size_20, test/test_nestedtensor.py::TestNestedTensor::test_3d_nested_tensor_float_batch_size_2_max_seq_len_3_vocab_size_10, test/test_nestedtensor.py::TestNestedTensor::test_cat, test/test_nestedtensor.py::TestNestedTensor::test_is_contiguous, test/test_nestedtensor.py::TestNestedTensor::test_nested_view_from_buffer_overflow_errors, test/test_nestedtensor.py::TestNestedTensor::test_numel, test/test_nestedtensor.py::TestNestedTensor::test_repr_string, test/test_nestedtensor.py::TestNestedTensor::test_size_dim, test/test_nestedtensor.py::TestNestedTensor::test_stride, test/test_nestedtensor.py::TestNestedTensor::test_to, test/test_nestedtensor.py::TestNestedTensor::test_to_padded_tensor_on_empty_tensor, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_bmm_cpu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_bmm_cpu_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_bmm_cuda_cuda_bfloat16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_bmm_noncontiguous_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_clone_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_contiguous_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_dropout_jagged_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_empty_like_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_layer_norm_breaking_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_linear_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_masked_fill_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_matmul_noncontiguous_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_matmul_noncontiguous_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_matmul_nt_with_broadcasted_t_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_narrow_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_add_in_place_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_chunk_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_dense_elementwise_embedding_dim_256_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_dense_elementwise_embedding_dim_256_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_dense_elementwise_embedding_dim_384_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_dense_elementwise_embedding_dim_8_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_div_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_div_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_indexing_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_indexing_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_mul_in_place_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_sub_transpose_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_sub_transpose_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_sub_transpose_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_sub_transpose_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_nested_tensor_sum_dim_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_reshape_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_serialization_requires_grad_False_weights_only_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_serialization_requires_grad_False_weights_only_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_serialization_requires_grad_True_weights_only_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_serialization_requires_grad_True_weights_only_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_serialization_requires_grad_True_weights_only_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_softmax_noncontiguous_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_squeeze_unsqueeze_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_dim2_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_dim2_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_dim3_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_dim3_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_dim4_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_output_size_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_simple_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_zero_numel_errors_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_padded_tensor_zero_numel_errors_cuda_float64, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_to_then_from_padded_tensor_no_transform0213_cuda_float32, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_transpose_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_transpose_inference_mode_interaction_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_unary_funcs_gelu_cuda, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_unary_funcs_isneginf_cuda, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_unary_funcs_logical_not_cuda, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_unary_funcs_neg_cuda, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_view_cuda_float16, test/test_nestedtensor.py::TestNestedTensorDeviceTypeCUDA::test_view_inference_mode_interaction_cuda_float16, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_accumulate_grad_different_strides_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_as_nested_tensor_propagates_gradients_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_backward_add_strided_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_backward_for_add_op_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_backward_sub_strided_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_gelu_backward_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_indexing_backward_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_layer_norm_backward_5d_size_32_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_layer_norm_backward_size_1024_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_layer_norm_backward_size_512_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_nested_tensor_from_list_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_nested_tensor_from_mask_and_to_padded_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_nested_tensor_from_padded_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_nested_tensor_matmul_gradcheck_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_nested_tensor_squeeze_gradcheck_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_nested_tensor_transpose_backward_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_nested_tensor_transpose_gradcheck_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_selu_backward_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_set_requires_grad_from_list_cuda, test/test_nestedtensor.py::TestNestedTensorAutogradCUDA::test_to_buffer_series_ops_grad_with_broadcast_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_jagged_requires_grad_False_contiguous_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_jagged_requires_grad_False_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_jagged_requires_grad_True_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_jagged_requires_grad_True_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_jagged_requires_grad_True_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_strided_requires_grad_False_contiguous_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_strided_requires_grad_False_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_strided_requires_grad_False_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_strided_requires_grad_True_contiguous_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_strided_requires_grad_True_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_0_layout_strided_requires_grad_True_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_jagged_requires_grad_False_contiguous_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_jagged_requires_grad_False_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_jagged_requires_grad_False_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_jagged_requires_grad_False_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_jagged_requires_grad_True_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_strided_requires_grad_False_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_strided_requires_grad_False_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_strided_requires_grad_True_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_strided_requires_grad_True_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_1_layout_strided_requires_grad_True_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_jagged_requires_grad_False_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_jagged_requires_grad_False_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_jagged_requires_grad_True_contiguous_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_jagged_requires_grad_True_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_strided_requires_grad_False_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_2_layout_strided_requires_grad_True_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_jagged_requires_grad_False_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_jagged_requires_grad_False_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_jagged_requires_grad_False_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_jagged_requires_grad_True_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_jagged_requires_grad_True_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_strided_requires_grad_False_contiguous_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_strided_requires_grad_False_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_strided_requires_grad_False_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_strided_requires_grad_True_contiguous_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_strided_requires_grad_True_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_3_layout_strided_requires_grad_True_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_4_layout_jagged_requires_grad_False_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_4_layout_jagged_requires_grad_False_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_4_layout_jagged_requires_grad_True_contiguous_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_4_layout_jagged_requires_grad_True_contiguous_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_4_layout_jagged_requires_grad_True_contiguous_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_4_layout_strided_requires_grad_False_contiguous_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_4_layout_strided_requires_grad_False_contiguous_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_as_nested_tensor_from_tensor_dim_4_layout_strided_requires_grad_True_contiguous_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_binary_pointwise_broadcasting_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_binary_pointwise_transposed_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_compile_padded_dense_conversion_preserves_metadata_cache_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_composite_op_in_inference_mode_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_construction_from_list_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_copy__cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_device_dtype_transfer_updates_offsets_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_dropout_inference_mode_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_dummy_mha_with_nt_use_legacy_api_False_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_flex_attention_noncontig_with_holes_False_cross_attention_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_flex_attention_noncontig_with_holes_False_cross_attention_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_as_nested_tensor_components_require_grad_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_nested_tensor_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_nested_tensor_requires_grad_True_components_require_grad_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_nested_tensor_requires_grad_True_components_require_grad_True_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_nested_tensor_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_layout_construction_with_pinned_memory_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_op_different_output_shape_dim_mean_keepdim_False_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_op_different_output_shape_dim_mean_keepdim_True_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_op_different_output_shape_dim_mean_keepdim_True_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_op_different_output_shape_dim_mean_keepdim_True_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_op_different_output_shape_dim_sum_keepdim_True_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_view_from_values_offsets_requires_grad_False_values_is_view_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_view_from_values_offsets_requires_grad_False_values_is_view_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_view_from_values_offsets_requires_grad_True_values_is_view_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_jagged_view_from_values_offsets_requires_grad_True_values_is_view_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_layer_norm_2d_input_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_layer_norm_2d_input_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_layer_norm_operate_on_batch_dim_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_layer_norm_operate_on_batch_dim_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_layer_norm_reduce_ragged_idx_1_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_like_shape_empty_like_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_like_shape_randn_like_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_like_value_rand_like_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_narrow_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_nested_tensor_from_jagged_pass_min_max_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_njt_cat_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_noncontiguous_to_noncontig_transposed_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_batch_only_different_output_shape_mean_keepdim_False_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_batch_only_different_output_shape_mean_keepdim_False_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_batch_only_different_output_shape_mean_keepdim_False_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_batch_only_different_output_shape_sum_keepdim_True_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_1_different_output_shape_mean_keepdim_False_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_1_different_output_shape_mean_keepdim_False_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_1_different_output_shape_mean_keepdim_False_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_1_different_output_shape_mean_keepdim_True_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_1_different_output_shape_sum_keepdim_False_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_1_different_output_shape_sum_keepdim_False_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_1_different_output_shape_sum_keepdim_False_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_1_different_output_shape_sum_keepdim_True_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_greater_than_1_different_output_shape_mean_transpose_offset_1_keepdim_False_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_greater_than_1_different_output_shape_mean_transpose_offset_2_keepdim_False_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_greater_than_1_different_output_shape_mean_transpose_offset_2_keepdim_False_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_greater_than_1_different_output_shape_sum_transpose_offset_1_keepdim_False_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_greater_than_1_different_output_shape_sum_transpose_offset_2_keepdim_True_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_reduce_ragged_idx_greater_than_1_different_output_shape_sum_transpose_offset_2_keepdim_True_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_transpose_non_ragged_dim_different_output_shape_mean_keepdim_False_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_transpose_non_ragged_dim_different_output_shape_mean_keepdim_False_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_transpose_non_ragged_dim_different_output_shape_mean_keepdim_False_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_transpose_non_ragged_dim_different_output_shape_mean_keepdim_True_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_transpose_non_ragged_dim_different_output_shape_mean_keepdim_True_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_transpose_non_ragged_dim_different_output_shape_sum_keepdim_True_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_transpose_non_ragged_dim_different_output_shape_sum_keepdim_True_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_transpose_non_ragged_dim_different_output_shape_sum_keepdim_True_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_with_lengths_different_output_shape_mean_keepdim_False_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_with_lengths_different_output_shape_mean_keepdim_False_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_with_lengths_different_output_shape_mean_keepdim_True_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_with_lengths_different_output_shape_mean_keepdim_True_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_with_lengths_different_output_shape_mean_keepdim_True_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_with_lengths_different_output_shape_sum_keepdim_False_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_with_lengths_different_output_shape_sum_keepdim_False_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_op_dim_with_lengths_different_output_shape_sum_keepdim_False_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_reshape_decomp_requires_grad_True_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_autocast_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_backwards_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_backwards_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_with_constant_sequence_length_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_with_constant_sequence_length_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_with_packed_in_proj_cuda_bfloat16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_with_packed_in_proj_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sdpa_with_packed_in_proj_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_serialization_noncontig_transposed_weights_only_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_serialization_noncontig_with_holes_weights_only_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_reduce_ragged_idx_1_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_reduce_ragged_idx_1_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_reduce_ragged_idx_greater_than_1_same_output_shape_transpose_offset_1_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_reduce_ragged_idx_greater_than_1_same_output_shape_transpose_offset_1_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_reduce_ragged_idx_greater_than_1_same_output_shape_transpose_offset_2_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_dim_with_lengths_requires_grad_False_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_softmax_reduce_batch_dim_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_split_with_sizes_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_squeeze_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sum_dim_reduce_batch_and_non_batch_keepdim_False_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sum_dim_reduce_batch_and_non_batch_keepdim_True_requires_grad_True_components_require_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sum_dim_reduce_ragged_and_non_batch_keepdim_False_requires_grad_False_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_sum_dim_reduce_ragged_and_non_batch_keepdim_True_requires_grad_True_components_require_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_threshold_backward_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_copy_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_dtype_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_2_requires_grad_False_cuda_float16, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_2_requires_grad_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_2_requires_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_2_requires_grad_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_3_requires_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_3_requires_grad_False_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_3_requires_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_3_requires_grad_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_4_requires_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_4_requires_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_compile_nt_dim_4_requires_grad_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_2_requires_grad_True_cuda_bool, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_2_requires_grad_True_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_3_requires_grad_True_cuda_bool, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_3_requires_grad_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_4_requires_grad_False_cuda_bool, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_4_requires_grad_False_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_to_padded_tensor_nt_dim_4_requires_grad_True_cuda_float64, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_unbind_backward_cuda_float32, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_unbind_lengths_ragged_idx_2_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_unbind_transpose_ragged_idx_2_cuda, test/test_nestedtensor.py::TestNestedTensorSubclassCUDA::test_view_ragged_idx_not_one_cuda, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward___rdiv___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward___rmod___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_acos_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_asin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_asinh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_atan2_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_atan_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_atanh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_cdouble_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_div_trunc_rounding_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_double_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_float_power_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_fmod_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_i0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_index_put_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_lgamma_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_logaddexp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_masked_amax_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_masked_amin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_masked_norm_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_masked_prod_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_masked_std_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_min_binary_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_min_reduction_with_dim_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nan_to_num_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_narrow_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nn_functional_prelu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nn_functional_rms_norm_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nn_functional_softsign_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_nn_functional_threshold_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_polygamma_polygamma_n_1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_pow_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_prod_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_round_decimals_0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_select_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_sigmoid_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_sinc_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_special_entr_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_special_i0e_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_special_xlog1py_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_sqrt_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_squeeze_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_std_unbiased_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_sum_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_tan_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_tanh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_unsqueeze_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_backward_xlogy_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward___rmul___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward___rpow___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_acos_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_amax_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_amin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_atanh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_ceil_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_clamp_min_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_erfinv_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_fmax_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_hypot_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_index_put_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_log2_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_masked_amin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_masked_logsumexp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_masked_select_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_max_binary_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_maximum_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_min_reduction_with_dim_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_minimum_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_mul_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nan_to_num_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nanmean_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_narrow_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nn_functional_celu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nn_functional_elu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nn_functional_linear_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nn_functional_mish_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nn_functional_relu6_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nn_functional_rms_norm_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nn_functional_rrelu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nn_functional_silu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_nn_functional_softshrink_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_remainder_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_rsqrt_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_sigmoid_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_sin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_sinh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_special_entr_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_special_i0e_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_special_i1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_special_log_ndtr_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_special_ndtr_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_split_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_split_with_sizes_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_sqrt_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_sum_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_tan_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_tanh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_to_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_unflatten_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_unsqueeze_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_backward_xlogy_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward___radd___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward___rdiv___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_argmax_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_argmin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_asin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_atanh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_bfloat16_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_byte_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_chalf_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_char_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_chunk_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_clamp_max_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_clamp_min_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_conj_physical_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_copysign_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_cosh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_deg2rad_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_digamma_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_div_floor_rounding_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_div_trunc_rounding_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_double_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_eq_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_erf_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_erfinv_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_expm1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_floor_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_floor_divide_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_fmax_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_frexp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_gt_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_i0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_index_put_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_isclose_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_isnan_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_jiterator_binary_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_jiterator_binary_return_by_ref_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_linalg_vector_norm_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_log_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_logaddexp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_long_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_masked_logsumexp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_masked_std_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_max_binary_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_mean_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_mvlgamma_mvlgamma_p_3_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nanmean_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_narrow_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nn_functional_hardsigmoid_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nn_functional_prelu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nn_functional_relu6_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nn_functional_rms_norm_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_nn_functional_threshold_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_polygamma_polygamma_n_0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_polygamma_polygamma_n_1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_polygamma_polygamma_n_2_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_polygamma_polygamma_n_4_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_pow_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_prod_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_remainder_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_round_decimals_3_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_round_decimals_neg_3_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_rsqrt_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_rsub_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_sgn_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_bessel_j1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_bessel_y1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_chebyshev_polynomial_w_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_hermite_polynomial_h_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_hermite_polynomial_he_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_i1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_i1e_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_modified_bessel_i0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_ndtri_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_scaled_modified_bessel_k1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_shifted_chebyshev_polynomial_t_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_split_with_sizes_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_square_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_tanh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_to_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_trunc_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_unflatten_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_var_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_compile_forward_where_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward___radd___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward___rdiv___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward___rmul___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward___rpow___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward___rsub___cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_argmax_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_asinh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_cdouble_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_chalf_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_chunk_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_clamp_max_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_clamp_min_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_count_nonzero_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_deg2rad_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_digamma_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_div_no_rounding_mode_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_erfinv_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_exp2_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_exp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_float_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_frac_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_frexp_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_hypot_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_igamma_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_igammac_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_index_put_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_isfinite_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_isposinf_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_jiterator_binary_return_by_ref_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_linalg_vector_norm_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_log10_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_log1p_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_log_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_logical_not_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_masked_amax_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_masked_argmin_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_masked_prod_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_masked_sum_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_max_binary_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_max_reduction_with_dim_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nan_to_num_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nanmean_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nn_functional_celu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nn_functional_elu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nn_functional_relu6_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nn_functional_rrelu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nn_functional_selu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nn_functional_silu_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nn_functional_softplus_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_nn_functional_softsign_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_positive_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_prod_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_rad2deg_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_rsub_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_sigmoid_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_sign_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_sinc_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_sinh_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_airy_ai_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_bessel_j1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_chebyshev_polynomial_v_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_hermite_polynomial_he_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_log_ndtr_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_modified_bessel_i0_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_modified_bessel_i1_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_ndtri_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_special_shifted_chebyshev_polynomial_t_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_std_unbiased_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_sum_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_to_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_var_unbiased_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_forward_where_cuda_float32, test/test_nestedtensor.py::TestNestedTensorOpInfoCUDA::test_nested_tensor_non_contiguous_mutation_cuda 2025-07-17T09:09:36.2004536Z 2025-07-17T09:09:36.2004737Z Running test_transformers 1/1 ... [2025-07-17 09:09:36.169167] 2025-07-17T09:09:36.2005270Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:09:36.2005583Z GITHUB_RUN_ID, GITHUB_RUN_ATTEMPT, or ARTIFACTS_FILE_SUFFIX not set, not uploading 2025-07-17T09:09:36.2005991Z Uploading artifacts took 0.00 seconds 2025-07-17T09:09:36.2006739Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_transformers.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:09:36.169475] 2025-07-17T09:15:07.7185827Z 2025-07-17T09:15:07.7186652Z inductor/test_torchinductor_dynamic_shapes 2/6 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_dynamic_shapes_2.6_63e4b711bdad3cb6_.log 2025-07-17T09:15:07.7278116Z Running 320 items in this shard: test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_add_complex5_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_aoti_eager_override_registration_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_aoti_eager_with_scalar_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_arange1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_as_strided_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_assert_alignment_op_name_fail_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_assert_size_stride_op_name_fail_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_assert_size_stride_op_name_pass_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_int_uint8_int16_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_bucketize_nd_tiling_True_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_buffer_batch_norm_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_builtins_round_float_ndigits_neg_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cat_unbacked_legacy_empty_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cauchy_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_check_stack_no_cycles_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_clamp_type_promotion_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_complex_fallback_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_computed_buffer_inlining_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_concat_add_inplace_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_config_option_dont_assume_alignment_cudagraphs_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_const_int32_to_float_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_convolution1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_convolution4_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cos_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cumsum_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cumsum_inf_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cumsum_pattern_matcher_issue_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_cumsum_zero_dim_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_div_by_zero_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_div_prim_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dropout2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dropout3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dropout_trivial_1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_bfloat16_float32_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_bfloat16_float64_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_bfloat16_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float16_float64_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float32_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float32_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float64_bfloat16_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_float64_float32_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int32_bfloat16_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int32_float32_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int32_int32_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int32_uint8_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int64_float16_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int64_float32_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int64_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_int8_uint8_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_dtypeview_uint8_int64_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_elu_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_exact_stride_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_exp2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_expand_as_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_expand_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_flip_cat_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_float_repr_dynamic_shapes_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_forced_buffer_realize_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_fractional_max_pool2d3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_fuse_large_params_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_gather_scatter_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_graph_partition_both_scalars_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_index_propagation_abs_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inductor_assert_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inf_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inner_fn_str_and_stride_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inplace_add_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inplace_mixed_dtype_ops_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_inplace_where_pointwise_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_int8_weight_only_quant_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_isinf_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_large_grid_use_block_ptr_False_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_linspace4_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_low_memory_max_pool_dilation_1_dim_3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_matmul_layer_norm_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_max_pool2d2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_max_pool2d8_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_max_pool2d_with_indices_backward_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mean_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_min_max_reduction_nan_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mixed_mm_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_move_arange_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_multi_device_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_multi_gpu_recompile_on_index_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_mutable_custom_op_fixed_layout2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_nan_to_num_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_narrow_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_neg_max_uint8_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_new_empty_strided_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_nll_loss_backward_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_nonzero_unbacked_refinement_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pixel_shuffle_channels_last_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_bessel_y1_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_chebyshev_polynomial_t_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_chebyshev_polynomial_w_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_entr_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_gammaln_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_i0e_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_modified_bessel_k0_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_psi_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_scaled_modified_bessel_k0_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_pointwise_shifted_chebyshev_polynomial_v_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_prod_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_randint_int64_mod_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_randint_kernel_count_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_reduction3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_reduction5_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_reflection_pad2d_backward_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_reflection_pad2d_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_remove_noop_slice_scatter_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_repeat_as_strided_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_resize_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_roi_align_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_searchsorted_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_setitem_with_int_parameter_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_single_elem_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_slice3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_slice_mutation3_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_slice_scatter4_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_slice_view_with_graph_break_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_softmax_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sort_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sum2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_sum5_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_tan_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_tensor_index_put_slice_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_tmp_not_defined_issue2_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_transpose_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_uint4x2_mixed_mm_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_unbacked_floordiv_simplify_errors_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_unbind_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_unspec_inputs_float16_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_unspec_inputs_float64_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_var_correction_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_var_mean_div_by_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_view_as_real_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_views6_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_where_with_logical_op_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesCpuTests::test_zero_element_mutation_dynamic_shapes_cpu, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_abs_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_adaptive_avg_pool2d_low_prec_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_adaptive_avg_pool_with_output_size_0_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_aliased_buffer_reuse_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_aoti_eager_cache_hit_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_arange2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_arange6_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_argmax_argmin2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_argmax_argmin3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_argmax_to_float_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_as_strided_scatter_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_assert_size_stride_op_name_pass_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_batch_norm_2d_2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_computed_offsets_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_default_kwargs_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_int16_int16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_int16_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_int8_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_int8_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_uint8_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_int_uint8_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_bucketize_nd_tiling_True_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_builtins_round_float_ndigits_pos_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_builtins_round_int_ndigits_pos_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_builtins_round_int_ndigits_zero_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_cat_inplace_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_clone_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_config_option_dont_assume_alignment_cudagraphs_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_config_option_dont_assume_alignment_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_convolution1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_convolution5_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_custom_op_default_layout_constraint_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_custom_op_fixed_layout_channels_last_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_custom_scan_op_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_div2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_div4_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_div8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_div9_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_div_by_zero_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_div_precision_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dropout3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dropout_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtype_mismatch_issue_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_bfloat16_float16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float16_float16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float16_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float32_int16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float32_int64_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_float32_uint8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int32_float64_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int64_float32_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int64_int16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int64_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int8_float32_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int8_int16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_int8_int32_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_uint8_float16_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_dtypeview_uint8_int8_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_erfinv_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_expanded_reduction_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_fallback_mutable_op_basic_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_fill2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_flip_cat_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_flip_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_float_index_expression_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_float_repr_dynamic_shapes_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_fractional_max_pool2d2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_fractional_max_pool2d5_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_full_boolean_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_full_truncation_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_gather1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_gather2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_getitem_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_graph_partition_arange1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_graph_partition_constant_tensor1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_graph_partition_unbacked_symint_as_output_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_index_propagation_nested_indirect_indexing_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_index_put1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_index_put2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_index_put_fallback2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_index_select_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_index_tensor_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_inductor_layout_optimization_input_mutations_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_inner_fn_str_and_stride_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_inplace_add_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_input_mutation5_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_issue102546_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_kwargs_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_large_strided_reduction_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_leaky_relu_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_linspace4_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_log1p_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_logaddexp_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_low_memory_max_pool_dilation_1_dim_3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_masked_fill_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_max_pool2d_with_indices_backward3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_mean_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_mm_mixed_dtype_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_move_arange_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_multilayer_prime_size_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_mutable_custom_op_fixed_layout2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_neg_index_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_no_mega_fusion_during_lowering_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_no_op_reduction_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_no_specization_over_symbolic_value_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_nonzero_unbacked_refinement_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pad_view_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pattern_matcher_multi_user_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_permute1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_chebyshev_polynomial_w_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_erf_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_gammaincc_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_gammaln_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_hermite_polynomial_h_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_laguerre_polynomial_l_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_scaled_modified_bessel_k0_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_shifted_chebyshev_polynomial_t_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_shifted_chebyshev_polynomial_u_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pointwise_zeta_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pow_by_natural_log2_dynamic_shapes_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_pow_int_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_reduction4_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_reduction5_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_remove_noop_clone_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_remove_noop_slice1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_remove_noop_slice_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_repeat_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_require_stride_expanded_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_roll_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_rsqrt_dynamic_shapes_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scalar_output_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scaled_dot_product_efficient_attention_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scatter1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scatter2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scatter_add2_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_scatter_add3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sdpa_unaligned_mask_freezing_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_searchsorted_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_select_scatter_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_shape_padding_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_shape_prop_torch_ones_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_signbit_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_single_elem_indirect_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_slice_scatter4_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_slice_scatter_reinplace_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sort_stable_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_split_failed_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_split_with_integer_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_split_with_list_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_stride_preservation_with_stride_modifying_fx_pass_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_sum1_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_tanh_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_tensor3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_tensor_index_slice_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_tmp_not_defined_issue1_use_block_ptr_True_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_tmp_not_defined_issue3_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_unsqueeze_inplace_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_upsample_bilinear2d_b_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_var_correction_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_view_uint8_through_differing_bitwidths_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_views6_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_views7_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_where_broadcast_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::DynamicShapesGPUTests::test_zero_dim_reductions_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_abs_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_arange_dynamic_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_cat_unbacked_duplicate_size_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_float_item_inf_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_float_item_return_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_item_zeros_nobreak_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_math_ops_op10_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_math_ops_op4_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_math_ops_op9_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_meta_dynamic_shapes_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_multi_output_unbacked_custom_op_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_noops_tensor_repropagate_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_return_unbacked_view_split_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_slice_scatter_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_sym_stride_lowering_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_symint_sum_list_cuda, test/inductor/test_torchinductor_dynamic_shapes.py::TestInductorDynamicCUDA::test_unbacked_cat_backwards_cuda 2025-07-17T09:15:07.7367789Z 2025-07-17T09:15:07.7367910Z Running test_decomp 2/12 ... [2025-07-17 09:15:07.718774] 2025-07-17T09:15:07.7368162Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:15:07.7368798Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_decomp.py', '-m', 'not serial', '--shard-id=2', '--num-shards=12', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:15:07.719074] 2025-07-17T09:17:06.7860315Z 2025-07-17T09:17:06.7863666Z PRINTING LOG FILE of test_transformers 1/1 (test/test-reports/test_transformers_1.1_228e957a862f69ea_.log) 2025-07-17T09:17:06.7879743Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T09:17:06.7880647Z import pkg_resources 2025-07-17T09:17:06.7881012Z Test results will be stored in test-reports/python-pytest/test_transformers/test_transformers-c3d3be2371ce67e4.xml 2025-07-17T09:17:06.7881411Z ============================= test session starts ============================== 2025-07-17T09:17:06.7881773Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T09:17:06.7882097Z cachedir: .pytest_cache 2025-07-17T09:17:06.7882468Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T09:17:06.7882868Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T09:17:06.7883065Z configfile: pytest.ini 2025-07-17T09:17:06.7883442Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T09:17:06.7884343Z collecting ... collected 12242 items 2025-07-17T09:17:06.7884815Z stepcurrent: Cannot find last run test, not skipping 2025-07-17T09:17:07.3671522Z Running 12242 items in this shard: test/test_transformers.py::TestTransformersCUDA::test_bias_is_none_cuda, test/test_transformers.py::TestTransformersCUDA::test_decoder_only_layer_cuda, test/test_transformers.py::TestTransformersCUDA::test_decoder_padding_and_src_mask_bool_cuda, test/test_transformers.py::TestTransformersCUDA::test_disable_fastpath_cuda, test/test_transformers.py::TestTransformersCUDA::test_encoder_is_causal_cuda, test/test_transformers.py::TestTransformersCUDA::test_encoder_padding_and_src_mask_bool_cuda, test/test_transformers.py::TestTransformersCUDA::test_is_causal_gpu_cuda, test/test_transformers.py::TestTransformersCUDA::test_kpm_mask_trailing_column_with_nested_tensor_cuda, test/test_transformers.py::TestTransformersCUDA::test_mask_check_fastpath_cuda, test/test_transformers.py::TestTransformersCUDA::test_math_backend_high_precision_cuda, test/test_transformers.py::TestTransformersCUDA::test_mha_native_args_nb_heads_1_bias_False_cuda, test/test_transformers.py::TestTransformersCUDA::test_mha_native_args_nb_heads_1_bias_True_cuda, test/test_transformers.py::TestTransformersCUDA::test_mha_native_args_nb_heads_8_bias_False_cuda, test/test_transformers.py::TestTransformersCUDA::test_mha_native_args_nb_heads_8_bias_True_cuda, test/test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim2_key_padding_mask_dim1_bool_cuda, test/test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim2_key_padding_mask_dim1_float32_cuda, test/test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim2_key_padding_mask_dim_2_bool_cuda, test/test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim2_key_padding_mask_dim_2_float32_cuda, test/test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_2_key_padding_mask_dim1_bool_cuda, test/test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_2_key_padding_mask_dim1_float32_cuda, test/test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_2_key_padding_mask_dim_2_bool_cuda, test/test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_2_key_padding_mask_dim_2_float32_cuda, test/test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_3_key_padding_mask_dim1_bool_cuda, test/test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_3_key_padding_mask_dim1_float32_cuda, test/test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_3_key_padding_mask_dim_2_bool_cuda, test/test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_3_key_padding_mask_dim_2_float32_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_2D_attn_mask_dropout_p_0_0_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_2D_attn_mask_dropout_p_0_2_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_2D_attn_mask_dropout_p_0_5_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_2D_causal_attn_mask_dropout_p_0_0_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_2D_causal_attn_mask_dropout_p_0_2_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_2D_causal_attn_mask_dropout_p_0_5_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_3D_attn_mask_dropout_p_0_0_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_3D_attn_mask_dropout_p_0_2_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_3D_attn_mask_dropout_p_0_5_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_3D_causal_attn_mask_dropout_p_0_0_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_3D_causal_attn_mask_dropout_p_0_2_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_3D_causal_attn_mask_dropout_p_0_5_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_no_attn_mask_dropout_p_0_0_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_no_attn_mask_dropout_p_0_2_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_no_attn_mask_dropout_p_0_5_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_2D_attn_mask_dropout_p_0_0_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_2D_attn_mask_dropout_p_0_2_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_2D_attn_mask_dropout_p_0_5_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_2D_causal_attn_mask_dropout_p_0_0_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_2D_causal_attn_mask_dropout_p_0_2_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_2D_causal_attn_mask_dropout_p_0_5_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_4D_attn_mask_dropout_p_0_0_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_4D_attn_mask_dropout_p_0_2_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_4D_attn_mask_dropout_p_0_5_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_4D_causal_attn_mask_dropout_p_0_0_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_4D_causal_attn_mask_dropout_p_0_2_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_4D_causal_attn_mask_dropout_p_0_5_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_no_attn_mask_dropout_p_0_0_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_no_attn_mask_dropout_p_0_2_cuda, test/test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_no_attn_mask_dropout_p_0_5_cuda, test/test_transformers.py::TestTransformersCUDA::test_script_encoder_subclass_cuda, test/test_transformers.py::TestTransformersCUDA::test_script_mha_in_proj_weight_none_cuda, test/test_transformers.py::TestTransformersCUDA::test_self_attn_TxT_attn_mask_cuda, test/test_transformers.py::TestTransformersCUDA::test_train_with_is_causal_cuda, test/test_transformers.py::TestTransformersCUDA::test_train_with_pad_and_catch_error_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformer_bias_is_none_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_False_training_False_enable_nested_tensor_False_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_False_training_False_enable_nested_tensor_True_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_False_training_True_enable_nested_tensor_False_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_False_training_True_enable_nested_tensor_True_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_True_training_False_enable_nested_tensor_False_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_True_training_False_enable_nested_tensor_True_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_True_training_True_enable_nested_tensor_False_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_True_training_True_enable_nested_tensor_True_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_False_use_autocast_False_d_model_12_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_False_use_autocast_False_d_model_256_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_False_use_autocast_True_d_model_12_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_False_use_autocast_True_d_model_256_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_True_use_autocast_False_d_model_12_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_True_use_autocast_False_d_model_256_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_True_use_autocast_True_d_model_12_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_True_use_autocast_True_d_model_256_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_square_input_with_no_grad_False_training_False_enable_nested_tensor_False_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_square_input_with_no_grad_False_training_True_enable_nested_tensor_False_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_square_input_with_no_grad_True_training_False_enable_nested_tensor_False_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoder_square_input_with_no_grad_True_training_True_enable_nested_tensor_False_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_no_fastpath_with_hooks_nhead_3_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_no_fastpath_with_hooks_nhead_4_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_src_mask_nhead_1_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_src_mask_nhead_4_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_src_mask_nhead_8_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_subclass_cuda, test/test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_subclass_model_cuda, test/test_transformers.py::TestTransformersCUDA::test_with_nested_tensor_input_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_dispatch_fails_no_backend_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_flash_atteention_large_bf16_nan_values_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_flash_attention_fail_with_non_square_causal_attention_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_flash_autocast_fp32_bfloat16_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_flash_autocast_fp32_float16_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_flash_backward_failure_sm86plus_head_dim_193_dropout_p_0_0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_flash_backward_failure_sm86plus_head_dim_193_dropout_p_0_2_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_flash_backward_failure_sm86plus_head_dim_256_dropout_p_0_0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_flash_backward_failure_sm86plus_head_dim_256_dropout_p_0_2_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_flash_fail_fp32_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_fused_kernels_nested_broadcasting_error_cases_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_fused_kernels_nested_broadcasting_requires_grad_failure_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_fused_kernels_seq_len_0_inputs_fused_kernel0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_fused_kernels_seq_len_0_inputs_fused_kernel1_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_attn_mask_present_kernel0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_broadcast_kernel0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_broadcast_kernel1_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_dim_3_kernel0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_dim_3_kernel1_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_head_dim_kernel0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_head_dim_kernel1_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_invalid_dtype_kernel0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_invalid_dtype_kernel1_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_1_dimensional_inputs_kernel0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_1_dimensional_inputs_kernel1_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_1_dimensional_inputs_kernel2_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_different_datatypes_kernel0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_different_datatypes_kernel1_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_different_datatypes_kernel2_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_different_devices_kernel0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_different_devices_kernel1_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_different_devices_kernel2_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_last_dim_stride_kernel0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_last_dim_stride_kernel1_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_sdpa_kernel_grouped_query_attention_cuda_fused_kernel0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_sequence_lengths_kernel0_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_sequence_lengths_kernel1_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_mem_eff_attention_fail_with_batch_size_geq_65536_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_mem_eff_attention_fail_with_batch_size_geq_65536_error_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_mem_eff_attention_large_seq_len_uniform_attention_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_mem_efficient_fail_bfloat16_less_than_sm80_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_nested_fails_on_padding_head_dim_cuda, test/test_transformers.py::TestSDPAFailureModesCUDA::test_unaligned_tensors_cuda, test/test_transformers.py::TestSDPACUDA::test_scaled_dot_product_attention_math_with_negative_scale_kernel0_cuda, test/test_transformers.py::TestSDPACUDA::test_sdp_math_gradcheck_contiguous_inputs_False_cuda, test/test_transformers.py::TestSDPACUDA::test_sdp_math_gradcheck_contiguous_inputs_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_d256_heuristic_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_different_dk_dv_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_fail_d128_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_gqa_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_nonmodulo64seqlen_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_preserves_query_layout_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_trivial_output_transpose_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_different_dk_dv_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_backwards_throws_determinism_warning_fused_kernel0_warn_only_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_backwards_throws_determinism_warning_fused_kernel0_warn_only_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_backwards_throws_determinism_warning_fused_kernel1_warn_only_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_backwards_throws_determinism_warning_fused_kernel1_warn_only_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_query_dense_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_seq_len_1_inputs_fused_kernel0_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_seq_len_1_inputs_fused_kernel1_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_sdp_choice_type_dense_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_sdp_choice_type_nested_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_sdp_priority_order_use_compile_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_sdp_priority_order_use_compile_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_eff_attention_long_sequence_mask_float16_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_eff_attention_long_sequence_mask_float32_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_eff_attention_non_contig_mask_bug_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_eff_attention_non_contiguous_mask_float16_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_eff_attention_non_contiguous_mask_float32_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_eff_backwards_determinism_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_mask_variants_mask_dim_1_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_mask_variants_mask_dim_2_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_mask_variants_mask_dim_3_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_mask_variants_mask_dim_4_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_cudnn_nested_type_nested_is_contiguous_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_dense_fused_kernel0_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_dense_fused_kernel1_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_nested_fused_kernel0_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_nested_fused_kernel1_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_type_dense_is_contiguous_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_type_dense_is_contiguous_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_type_nested_is_contiguous_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_type_nested_is_contiguous_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_choice_with_determinism_warn_only_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_choice_with_determinism_warn_only_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_False_bfloat16_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_False_float16_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_True_bfloat16_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_True_float16_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_False_bfloat16_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_False_float16_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_True_bfloat16_cuda_bfloat16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_True_float16_cuda_float16, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_False_is_causal_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_False_is_causal_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_True_is_causal_False_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_True_is_causal_True_cuda, test/test_transformers.py::TestSDPACudaOnlyCUDA::test_singelton_head_dim_stride_ne_1_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_1_shape0_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_1_shape1_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_1_shape2_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_1_shape3_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_2_shape0_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_2_shape1_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_2_shape2_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_2_shape3_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_1_shape0_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_1_shape1_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_1_shape2_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_1_shape3_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_2_shape0_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_2_shape1_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_2_shape2_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_2_shape3_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_is_causal_and_mask_fails_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_is_causal_equals_upper_left_shape0_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_is_causal_equals_upper_left_shape1_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_is_causal_equals_upper_left_shape2_cuda, test/test_transformers.py::TestAttnBiasCUDA::test_is_causal_equals_upper_left_shape3_cuda 2025-07-17T09:17:07.9204658Z 2025-07-17T09:17:07.9204952Z test_transformers.py::TestTransformersCUDA::test_bias_is_none_cuda PASSED [0.0111s] [ 0%] 2025-07-17T09:17:07.9205494Z test_transformers.py::TestTransformersCUDA::test_decoder_only_layer_cuda SKIPPED [0.0002s] (Fairseq not found) [ 0%] 2025-07-17T09:17:07.9206035Z test_transformers.py::TestTransformersCUDA::test_decoder_padding_and_src_mask_bool_cuda PASSED [0.0089s] [ 0%] 2025-07-17T09:17:07.9206528Z test_transformers.py::TestTransformersCUDA::test_disable_fastpath_cuda PASSED [0.9616s] [ 0%] 2025-07-17T09:17:07.9206984Z test_transformers.py::TestTransformersCUDA::test_encoder_is_causal_cuda PASSED [0.0068s] [ 0%] 2025-07-17T09:17:07.9208573Z test_transformers.py::TestTransformersCUDA::test_encoder_padding_and_src_mask_bool_cuda PASSED [0.0045s] [ 0%] 2025-07-17T09:17:07.9209193Z test_transformers.py::TestTransformersCUDA::test_is_causal_gpu_cuda SKIPPED [0.0008s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 0%] 2025-07-17T09:17:07.9209862Z test_transformers.py::TestTransformersCUDA::test_kpm_mask_trailing_column_with_nested_tensor_cuda PASSED [0.0831s] [ 0%] 2025-07-17T09:17:07.9210388Z test_transformers.py::TestTransformersCUDA::test_mask_check_fastpath_cuda PASSED [0.0077s] [ 0%] 2025-07-17T09:17:07.9210865Z test_transformers.py::TestTransformersCUDA::test_math_backend_high_precision_cuda PASSED [7.0775s] [ 0%] 2025-07-17T09:17:07.9211375Z test_transformers.py::TestTransformersCUDA::test_mha_native_args_nb_heads_1_bias_False_cuda PASSED [0.0036s] [ 0%] 2025-07-17T09:17:07.9212017Z test_transformers.py::TestTransformersCUDA::test_mha_native_args_nb_heads_1_bias_True_cuda PASSED [0.0031s] [ 0%] 2025-07-17T09:17:07.9212536Z test_transformers.py::TestTransformersCUDA::test_mha_native_args_nb_heads_8_bias_False_cuda PASSED [0.0027s] [ 0%] 2025-07-17T09:17:07.9213046Z test_transformers.py::TestTransformersCUDA::test_mha_native_args_nb_heads_8_bias_True_cuda PASSED [0.0029s] [ 0%] 2025-07-17T09:17:07.9213663Z test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim2_key_padding_mask_dim1_bool_cuda PASSED [4.3524s] [ 0%] 2025-07-17T09:17:07.9215040Z test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim2_key_padding_mask_dim1_float32_cuda SKIPPED [0.0006s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157060 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9217022Z test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim2_key_padding_mask_dim_2_bool_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157038 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9218383Z test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim2_key_padding_mask_dim_2_float32_cuda PASSED [0.0048s] [ 0%] 2025-07-17T09:17:07.9219214Z test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_2_key_padding_mask_dim1_bool_cuda SKIPPED [0.0010s] (boolean mask is not fully supported on ROCm yet.) [ 0%] 2025-07-17T09:17:07.9220036Z test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_2_key_padding_mask_dim1_float32_cuda PASSED [0.0042s] [ 0%] 2025-07-17T09:17:07.9221412Z test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_2_key_padding_mask_dim_2_bool_cuda SKIPPED [0.0009s] (boolean mask is not fully supported on ROCm yet.) [ 0%] 2025-07-17T09:17:07.9222441Z test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_2_key_padding_mask_dim_2_float32_cuda PASSED [0.0039s] [ 0%] 2025-07-17T09:17:07.9223367Z test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_3_key_padding_mask_dim1_bool_cuda SKIPPED [0.0008s] (boolean mask is not fully supported on ROCm yet.) [ 0%] 2025-07-17T09:17:07.9224197Z test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_3_key_padding_mask_dim1_float32_cuda PASSED [0.0038s] [ 0%] 2025-07-17T09:17:07.9225026Z test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_3_key_padding_mask_dim_2_bool_cuda SKIPPED [0.0008s] (boolean mask is not fully supported on ROCm yet.) [ 0%] 2025-07-17T09:17:07.9225851Z test_transformers.py::TestTransformersCUDA::test_multiheadattention_fastpath_attn_mask_attn_mask_dim_3_key_padding_mask_dim_2_float32_cuda PASSED [0.0039s] [ 0%] 2025-07-17T09:17:07.9226554Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_2D_attn_mask_dropout_p_0_0_cuda PASSED [0.0764s] [ 0%] 2025-07-17T09:17:07.9227196Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_2D_attn_mask_dropout_p_0_2_cuda PASSED [0.0048s] [ 0%] 2025-07-17T09:17:07.9227828Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_2D_attn_mask_dropout_p_0_5_cuda PASSED [0.0044s] [ 0%] 2025-07-17T09:17:07.9229079Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_2D_causal_attn_mask_dropout_p_0_0_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157061 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9230944Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_2D_causal_attn_mask_dropout_p_0_2_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157039 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9232768Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_2D_causal_attn_mask_dropout_p_0_5_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157091 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9234067Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_3D_attn_mask_dropout_p_0_0_cuda PASSED [0.0032s] [ 0%] 2025-07-17T09:17:07.9234710Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_3D_attn_mask_dropout_p_0_2_cuda PASSED [0.0032s] [ 0%] 2025-07-17T09:17:07.9235347Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_3D_attn_mask_dropout_p_0_5_cuda PASSED [0.0032s] [ 0%] 2025-07-17T09:17:07.9236655Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_3D_causal_attn_mask_dropout_p_0_0_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157062 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9238628Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_3D_causal_attn_mask_dropout_p_0_2_cuda SKIPPED [0.0005s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157040 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9240593Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_3D_causal_attn_mask_dropout_p_0_5_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157092 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9242422Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_no_attn_mask_dropout_p_0_0_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/131086 for platform(s) linux, rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9244248Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_no_attn_mask_dropout_p_0_2_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/131146 for platform(s) linux, rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9246063Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_3D_input_dim_no_attn_mask_dropout_p_0_5_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/131123 for platform(s) linux, rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9247292Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_2D_attn_mask_dropout_p_0_0_cuda PASSED [0.0045s] [ 0%] 2025-07-17T09:17:07.9247933Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_2D_attn_mask_dropout_p_0_2_cuda PASSED [0.0042s] [ 0%] 2025-07-17T09:17:07.9248568Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_2D_attn_mask_dropout_p_0_5_cuda PASSED [0.0041s] [ 0%] 2025-07-17T09:17:07.9249806Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_2D_causal_attn_mask_dropout_p_0_0_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157063 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9251627Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_2D_causal_attn_mask_dropout_p_0_2_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157041 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9253702Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_2D_causal_attn_mask_dropout_p_0_5_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157093 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9255034Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_4D_attn_mask_dropout_p_0_0_cuda PASSED [0.0031s] [ 0%] 2025-07-17T09:17:07.9255683Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_4D_attn_mask_dropout_p_0_2_cuda PASSED [0.0031s] [ 0%] 2025-07-17T09:17:07.9256316Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_4D_attn_mask_dropout_p_0_5_cuda PASSED [0.0026s] [ 0%] 2025-07-17T09:17:07.9257566Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_4D_causal_attn_mask_dropout_p_0_0_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157064 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9259387Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_4D_causal_attn_mask_dropout_p_0_2_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157042 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9261207Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_4D_causal_attn_mask_dropout_p_0_5_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157094 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9263053Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_no_attn_mask_dropout_p_0_0_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/129853 for platform(s) linux, rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9265026Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_no_attn_mask_dropout_p_0_2_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/131107 for platform(s) linux, rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9266845Z test_transformers.py::TestTransformersCUDA::test_scaled_dot_product_attention_4D_input_dim_no_attn_mask_dropout_p_0_5_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/131179 for platform(s) linux, rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9267991Z test_transformers.py::TestTransformersCUDA::test_script_encoder_subclass_cuda PASSED [0.1617s] [ 0%] 2025-07-17T09:17:07.9268572Z test_transformers.py::TestTransformersCUDA::test_script_mha_in_proj_weight_none_cuda PASSED [0.0143s] [ 0%] 2025-07-17T09:17:07.9269244Z test_transformers.py::TestTransformersCUDA::test_self_attn_TxT_attn_mask_cuda SKIPPED [0.0002s] (4D mask not supported yet - activate when 4D mask supported) [ 0%] 2025-07-17T09:17:07.9269905Z test_transformers.py::TestTransformersCUDA::test_train_with_is_causal_cuda PASSED [3.1509s] [ 0%] 2025-07-17T09:17:07.9270568Z test_transformers.py::TestTransformersCUDA::test_train_with_pad_and_catch_error_cuda SKIPPED [0.0014s] (test is slow; run with PYTORCH_TEST_WITH_SLOW to enable test) [ 0%] 2025-07-17T09:17:07.9271178Z test_transformers.py::TestTransformersCUDA::test_transformer_bias_is_none_cuda PASSED [0.0240s] [ 0%] 2025-07-17T09:17:07.9271785Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_False_training_False_enable_nested_tensor_False_cuda PASSED [0.0249s] [ 0%] 2025-07-17T09:17:07.9272505Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_False_training_False_enable_nested_tensor_True_cuda PASSED [0.0245s] [ 0%] 2025-07-17T09:17:07.9273223Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_False_training_True_enable_nested_tensor_False_cuda PASSED [0.0260s] [ 0%] 2025-07-17T09:17:07.9274030Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_False_training_True_enable_nested_tensor_True_cuda PASSED [0.0260s] [ 0%] 2025-07-17T09:17:07.9274744Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_True_training_False_enable_nested_tensor_False_cuda PASSED [0.0209s] [ 0%] 2025-07-17T09:17:07.9275455Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_True_training_False_enable_nested_tensor_True_cuda PASSED [0.0368s] [ 0%] 2025-07-17T09:17:07.9276160Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_True_training_True_enable_nested_tensor_False_cuda PASSED [0.0266s] [ 0%] 2025-07-17T09:17:07.9276869Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_batch_first_True_training_True_enable_nested_tensor_True_cuda PASSED [0.0271s] [ 0%] 2025-07-17T09:17:07.9277787Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_False_use_autocast_False_d_model_12_cuda SKIPPED [0.0009s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 0%] 2025-07-17T09:17:07.9278887Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_False_use_autocast_False_d_model_256_cuda SKIPPED [0.0008s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 0%] 2025-07-17T09:17:07.9279964Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_False_use_autocast_True_d_model_12_cuda SKIPPED [0.0007s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 0%] 2025-07-17T09:17:07.9281040Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_False_use_autocast_True_d_model_256_cuda SKIPPED [0.0007s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 0%] 2025-07-17T09:17:07.9282125Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_True_use_autocast_False_d_model_12_cuda SKIPPED [0.0007s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 0%] 2025-07-17T09:17:07.9283218Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_True_use_autocast_False_d_model_256_cuda SKIPPED [0.0008s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 0%] 2025-07-17T09:17:07.9284293Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_True_use_autocast_True_d_model_12_cuda SKIPPED [0.0007s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 0%] 2025-07-17T09:17:07.9285520Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_fastpath_use_torchscript_False_enable_nested_tensor_True_use_autocast_True_d_model_256_cuda SKIPPED [0.0007s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 0%] 2025-07-17T09:17:07.9287129Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_square_input_with_no_grad_False_training_False_enable_nested_tensor_False_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157127 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9289138Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_square_input_with_no_grad_False_training_True_enable_nested_tensor_False_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157065 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9291073Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_square_input_with_no_grad_True_training_False_enable_nested_tensor_False_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157095 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9292996Z test_transformers.py::TestTransformersCUDA::test_transformerencoder_square_input_with_no_grad_True_training_True_enable_nested_tensor_False_cuda SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157043 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 0%] 2025-07-17T09:17:07.9294281Z test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_no_fastpath_with_hooks_nhead_3_cuda PASSED [0.0023s] [ 0%] 2025-07-17T09:17:07.9294907Z test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_no_fastpath_with_hooks_nhead_4_cuda PASSED [0.0020s] [ 0%] 2025-07-17T09:17:07.9295496Z test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_src_mask_nhead_1_cuda PASSED [0.0077s] [ 0%] 2025-07-17T09:17:07.9296054Z test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_src_mask_nhead_4_cuda PASSED [0.0026s] [ 0%] 2025-07-17T09:17:07.9296605Z test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_src_mask_nhead_8_cuda PASSED [0.2609s] [ 0%] 2025-07-17T09:17:07.9297146Z test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_subclass_cuda PASSED [0.1824s] [ 0%] 2025-07-17T09:17:07.9297686Z test_transformers.py::TestTransformersCUDA::test_transformerencoderlayer_subclass_model_cuda PASSED [0.7706s] [ 0%] 2025-07-17T09:17:07.9298200Z test_transformers.py::TestTransformersCUDA::test_with_nested_tensor_input_cuda PASSED [0.0108s] [ 0%] 2025-07-17T09:17:07.9298690Z test_transformers.py::TestSDPAFailureModesCUDA::test_dispatch_fails_no_backend_cuda PASSED [0.0019s] [ 0%] 2025-07-17T09:17:07.9299223Z test_transformers.py::TestSDPAFailureModesCUDA::test_flash_atteention_large_bf16_nan_values_cuda PASSED [0.0054s] [ 0%] 2025-07-17T09:17:07.9299819Z test_transformers.py::TestSDPAFailureModesCUDA::test_flash_attention_fail_with_non_square_causal_attention_cuda PASSED [0.0042s] [ 0%] 2025-07-17T09:17:07.9300383Z test_transformers.py::TestSDPAFailureModesCUDA::test_flash_autocast_fp32_bfloat16_cuda PASSED [0.0056s] [ 0%] 2025-07-17T09:17:07.9300965Z test_transformers.py::TestSDPAFailureModesCUDA::test_flash_autocast_fp32_float16_cuda PASSED [0.0057s] [ 0%] 2025-07-17T09:17:07.9301711Z test_transformers.py::TestSDPAFailureModesCUDA::test_flash_backward_failure_sm86plus_head_dim_193_dropout_p_0_0_cuda SKIPPED [0.0001s] (Does not support fused SDPA or not SM86+ hardware) [ 0%] 2025-07-17T09:17:07.9302638Z test_transformers.py::TestSDPAFailureModesCUDA::test_flash_backward_failure_sm86plus_head_dim_193_dropout_p_0_2_cuda SKIPPED [0.0001s] (Does not support fused SDPA or not SM86+ hardware) [ 0%] 2025-07-17T09:17:07.9303531Z test_transformers.py::TestSDPAFailureModesCUDA::test_flash_backward_failure_sm86plus_head_dim_256_dropout_p_0_0_cuda SKIPPED [0.0001s] (Does not support fused SDPA or not SM86+ hardware) [ 0%] 2025-07-17T09:17:07.9304353Z test_transformers.py::TestSDPAFailureModesCUDA::test_flash_backward_failure_sm86plus_head_dim_256_dropout_p_0_2_cuda SKIPPED [0.0001s] (Does not support fused SDPA or not SM86+ hardware) [ 0%] 2025-07-17T09:17:07.9305000Z test_transformers.py::TestSDPAFailureModesCUDA::test_flash_fail_fp32_cuda PASSED [0.0032s] [ 0%] 2025-07-17T09:17:07.9305541Z test_transformers.py::TestSDPAFailureModesCUDA::test_fused_kernels_nested_broadcasting_error_cases_cuda PASSED [0.0020s] [ 0%] 2025-07-17T09:17:07.9306165Z test_transformers.py::TestSDPAFailureModesCUDA::test_fused_kernels_nested_broadcasting_requires_grad_failure_cuda PASSED [0.0028s] [ 0%] 2025-07-17T09:17:07.9306790Z test_transformers.py::TestSDPAFailureModesCUDA::test_fused_kernels_seq_len_0_inputs_fused_kernel0_cuda PASSED [0.0020s] [ 0%] 2025-07-17T09:17:07.9307356Z test_transformers.py::TestSDPAFailureModesCUDA::test_fused_kernels_seq_len_0_inputs_fused_kernel1_cuda PASSED [0.0019s] [ 0%] 2025-07-17T09:17:07.9307943Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_attn_mask_present_kernel0_cuda PASSED [0.0009s] [ 0%] 2025-07-17T09:17:07.9308515Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_broadcast_kernel0_cuda PASSED [0.0009s] [ 0%] 2025-07-17T09:17:07.9309074Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_broadcast_kernel1_cuda PASSED [0.0009s] [ 0%] 2025-07-17T09:17:07.9309615Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_dim_3_kernel0_cuda PASSED [0.0021s] [ 0%] 2025-07-17T09:17:07.9310143Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_dim_3_kernel1_cuda PASSED [0.0016s] [ 0%] 2025-07-17T09:17:07.9310680Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_head_dim_kernel0_cuda PASSED [0.0009s] [ 0%] 2025-07-17T09:17:07.9311222Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_head_dim_kernel1_cuda PASSED [0.0008s] [ 0%] 2025-07-17T09:17:07.9311783Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_invalid_dtype_kernel0_cuda PASSED [0.0015s] [ 0%] 2025-07-17T09:17:07.9312352Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_fused_inputs_invalid_dtype_kernel1_cuda PASSED [0.0008s] [ 0%] 2025-07-17T09:17:07.9312928Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_1_dimensional_inputs_kernel0_cuda PASSED [0.0014s] [ 0%] 2025-07-17T09:17:07.9313501Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_1_dimensional_inputs_kernel1_cuda PASSED [0.0008s] [ 0%] 2025-07-17T09:17:07.9314153Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_1_dimensional_inputs_kernel2_cuda PASSED [0.0014s] [ 0%] 2025-07-17T09:17:07.9314730Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_different_datatypes_kernel0_cuda PASSED [0.0008s] [ 0%] 2025-07-17T09:17:07.9315299Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_different_datatypes_kernel1_cuda PASSED [0.0008s] [ 0%] 2025-07-17T09:17:07.9315868Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_different_datatypes_kernel2_cuda PASSED [0.0009s] [ 0%] 2025-07-17T09:17:07.9316563Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_different_devices_kernel0_cuda PASSED [0.0008s] [ 0%] 2025-07-17T09:17:07.9317196Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_different_devices_kernel1_cuda PASSED [0.0008s] [ 0%] 2025-07-17T09:17:07.9317826Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_inputs_different_devices_kernel2_cuda PASSED [0.0007s] [ 0%] 2025-07-17T09:17:07.9318436Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_last_dim_stride_kernel0_cuda PASSED [0.0026s] [ 1%] 2025-07-17T09:17:07.9318957Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_last_dim_stride_kernel1_cuda PASSED [0.0016s] [ 1%] 2025-07-17T09:17:07.9319549Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_sdpa_kernel_grouped_query_attention_cuda_fused_kernel0_cuda PASSED [0.0017s] [ 1%] 2025-07-17T09:17:07.9320143Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_sequence_lengths_kernel0_cuda PASSED [0.0016s] [ 1%] 2025-07-17T09:17:07.9320672Z test_transformers.py::TestSDPAFailureModesCUDA::test_invalid_sequence_lengths_kernel1_cuda PASSED [0.0016s] [ 1%] 2025-07-17T09:17:07.9321239Z test_transformers.py::TestSDPAFailureModesCUDA::test_mem_eff_attention_fail_with_batch_size_geq_65536_cuda PASSED [0.0652s] [ 1%] 2025-07-17T09:17:07.9321841Z test_transformers.py::TestSDPAFailureModesCUDA::test_mem_eff_attention_fail_with_batch_size_geq_65536_error_cuda PASSED [0.0014s] [ 1%] 2025-07-17T09:17:07.9322453Z test_transformers.py::TestSDPAFailureModesCUDA::test_mem_eff_attention_large_seq_len_uniform_attention_cuda PASSED [0.3657s] [ 1%] 2025-07-17T09:17:07.9323191Z test_transformers.py::TestSDPAFailureModesCUDA::test_mem_efficient_fail_bfloat16_less_than_sm80_cuda SKIPPED [0.0002s] (Current platform does not support fused SDPA or is an SM80+ device.) [ 1%] 2025-07-17T09:17:07.9323880Z test_transformers.py::TestSDPAFailureModesCUDA::test_nested_fails_on_padding_head_dim_cuda PASSED [0.0049s] [ 1%] 2025-07-17T09:17:07.9324378Z test_transformers.py::TestSDPAFailureModesCUDA::test_unaligned_tensors_cuda PASSED [0.0009s] [ 1%] 2025-07-17T09:17:07.9324906Z test_transformers.py::TestSDPACUDA::test_scaled_dot_product_attention_math_with_negative_scale_kernel0_cuda PASSED [0.0018s] [ 1%] 2025-07-17T09:17:07.9325448Z test_transformers.py::TestSDPACUDA::test_sdp_math_gradcheck_contiguous_inputs_False_cuda PASSED [0.1097s] [ 1%] 2025-07-17T09:17:07.9325944Z test_transformers.py::TestSDPACUDA::test_sdp_math_gradcheck_contiguous_inputs_True_cuda PASSED [0.0099s] [ 1%] 2025-07-17T09:17:07.9326549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_d256_heuristic_cuda SKIPPED [0.0002s] (cuDNN Attention is not supported on this system) [ 1%] 2025-07-17T09:17:07.9327263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_different_dk_dv_cuda SKIPPED [0.0001s] (cuDNN Attention is not supported on this system) [ 1%] 2025-07-17T09:17:07.9327914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_fail_d128_cuda SKIPPED [0.0001s] (broken as of cuDNN 9.10) [ 1%] 2025-07-17T09:17:07.9328539Z test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_gqa_cuda SKIPPED [0.0001s] (cuDNN Attention is not supported on this system) [ 1%] 2025-07-17T09:17:07.9329238Z test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_nonmodulo64seqlen_cuda SKIPPED [0.0001s] (cudnn Attention is not supported on this system) [ 1%] 2025-07-17T09:17:07.9329980Z test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_preserves_query_layout_cuda SKIPPED [0.0001s] (cudnn Attention is not supported on this system) [ 1%] 2025-07-17T09:17:07.9330730Z test_transformers.py::TestSDPACudaOnlyCUDA::test_cudnn_attention_trivial_output_transpose_cuda SKIPPED [0.0001s] (cudnn Attention is not supported on this system) [ 1%] 2025-07-17T09:17:07.9331621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0148s] [ 1%] 2025-07-17T09:17:07.9332759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 1%] 2025-07-17T09:17:07.9333816Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0044s] [ 1%] 2025-07-17T09:17:07.9334889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 1%] 2025-07-17T09:17:07.9335887Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 1%] 2025-07-17T09:17:07.9336897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 1%] 2025-07-17T09:17:07.9337895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 1%] 2025-07-17T09:17:07.9338887Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 1%] 2025-07-17T09:17:07.9339876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0133s] [ 1%] 2025-07-17T09:17:07.9340866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 1%] 2025-07-17T09:17:07.9341845Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0042s] [ 1%] 2025-07-17T09:17:07.9342823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0039s] [ 1%] 2025-07-17T09:17:07.9343809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 1%] 2025-07-17T09:17:07.9344808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 1%] 2025-07-17T09:17:07.9345787Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0042s] [ 1%] 2025-07-17T09:17:07.9346763Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0042s] [ 1%] 2025-07-17T09:17:07.9348147Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 [W717 09:10:00.196367045 attention.cpp:916] Warning: Dropout mask should only be used for testing purposes. (function operator()) 2025-07-17T09:17:07.9349010Z PASSED [0.0164s] [ 1%] 2025-07-17T09:17:07.9349589Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 1%] 2025-07-17T09:17:07.9350597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0045s] [ 1%] 2025-07-17T09:17:07.9351599Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0043s] [ 1%] 2025-07-17T09:17:07.9352597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 1%] 2025-07-17T09:17:07.9353602Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 1%] 2025-07-17T09:17:07.9354677Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0044s] [ 1%] 2025-07-17T09:17:07.9355673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0046s] [ 1%] 2025-07-17T09:17:07.9356666Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0157s] [ 1%] 2025-07-17T09:17:07.9357649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0040s] [ 1%] 2025-07-17T09:17:07.9358637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0043s] [ 1%] 2025-07-17T09:17:07.9359621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0042s] [ 1%] 2025-07-17T09:17:07.9360615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 1%] 2025-07-17T09:17:07.9361604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 1%] 2025-07-17T09:17:07.9362762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0044s] [ 1%] 2025-07-17T09:17:07.9363822Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0043s] [ 1%] 2025-07-17T09:17:07.9364877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 1%] 2025-07-17T09:17:07.9365883Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:07.9366885Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 1%] 2025-07-17T09:17:07.9367876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 1%] 2025-07-17T09:17:07.9368871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 1%] 2025-07-17T09:17:07.9369864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 1%] 2025-07-17T09:17:07.9370864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 1%] 2025-07-17T09:17:07.9371857Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 1%] 2025-07-17T09:17:07.9372845Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 1%] 2025-07-17T09:17:07.9373834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:07.9374816Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 1%] 2025-07-17T09:17:07.9375797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0040s] [ 1%] 2025-07-17T09:17:07.9376779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 1%] 2025-07-17T09:17:07.9377910Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:07.9378960Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 1%] 2025-07-17T09:17:07.9379999Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0039s] [ 1%] 2025-07-17T09:17:07.9381126Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9382376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9383608Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9384834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9386058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9387291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9388531Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9389765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9390996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9392270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9393623Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9394964Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9396189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9397423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9398641Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9399852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9401078Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9402322Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9403550Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9404788Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9406029Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9407329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9408719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9410013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9411250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9412483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9413711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9414924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9416147Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9417375Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9418598Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9419823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9421057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9422361Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9423710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9424996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9426229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9427468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9428705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9429937Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9431158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9432377Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9433588Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9434889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9436153Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9437451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9438802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9440257Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:07.9441372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0159s] [ 1%] 2025-07-17T09:17:07.9442385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 1%] 2025-07-17T09:17:07.9443649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 1%] 2025-07-17T09:17:07.9444790Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 1%] 2025-07-17T09:17:07.9445911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:07.9447046Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 2%] 2025-07-17T09:17:07.9448185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 2%] 2025-07-17T09:17:07.9449272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 2%] 2025-07-17T09:17:07.9450428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0169s] [ 2%] 2025-07-17T09:17:07.9451522Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0045s] [ 2%] 2025-07-17T09:17:07.9452626Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0097s] [ 2%] 2025-07-17T09:17:07.9453737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0046s] [ 2%] 2025-07-17T09:17:07.9455168Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0048s] [ 2%] 2025-07-17T09:17:07.9456379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0082s] [ 2%] 2025-07-17T09:17:07.9457495Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0057s] [ 2%] 2025-07-17T09:17:07.9458675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0053s] [ 2%] 2025-07-17T09:17:07.9459817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0287s] [ 2%] 2025-07-17T09:17:07.9460938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0048s] [ 2%] 2025-07-17T09:17:07.9462060Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 2%] 2025-07-17T09:17:07.9463164Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 2%] 2025-07-17T09:17:07.9464277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 2%] 2025-07-17T09:17:07.9465435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 2%] 2025-07-17T09:17:07.9466634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 2%] 2025-07-17T09:17:07.9467752Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 2%] 2025-07-17T09:17:07.9468862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0134s] [ 2%] 2025-07-17T09:17:07.9469977Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0039s] [ 2%] 2025-07-17T09:17:07.9471179Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 2%] 2025-07-17T09:17:07.9472401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 2%] 2025-07-17T09:17:07.9473535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0038s] [ 2%] 2025-07-17T09:17:07.9474783Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 2%] 2025-07-17T09:17:07.9475889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 2%] 2025-07-17T09:17:07.9477075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 2%] 2025-07-17T09:17:07.9478167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 2%] 2025-07-17T09:17:07.9479265Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 2%] 2025-07-17T09:17:07.9480422Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 2%] 2025-07-17T09:17:07.9481529Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 2%] 2025-07-17T09:17:07.9482653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 2%] 2025-07-17T09:17:07.9483774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 2%] 2025-07-17T09:17:07.9484879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 2%] 2025-07-17T09:17:07.9486014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 2%] 2025-07-17T09:17:07.9487113Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 2%] 2025-07-17T09:17:07.9488333Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 2%] 2025-07-17T09:17:07.9489610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 2%] 2025-07-17T09:17:07.9490822Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0039s] [ 2%] 2025-07-17T09:17:07.9491933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 2%] 2025-07-17T09:17:07.9493046Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 2%] 2025-07-17T09:17:07.9494144Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 2%] 2025-07-17T09:17:07.9495267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 2%] 2025-07-17T09:17:07.9496525Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9497884Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9499288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9500696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9502074Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9503469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9504902Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9506390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9507778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9509147Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9510478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9511804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9513168Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9514677Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9516029Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9517350Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9518699Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9520085Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9521503Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9522962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9524393Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9525744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9527213Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9528564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9529903Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9531265Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9532599Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9533955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9535281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9536598Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9538056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9539497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9541006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9542350Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9543671Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9545040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9546420Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9547792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9549198Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9550522Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9551895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9553239Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9554804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9556294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9557670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9559050Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9560436Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9561764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:07.9563006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0149s] [ 2%] 2025-07-17T09:17:07.9564108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:07.9565248Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 2%] 2025-07-17T09:17:07.9566427Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 2%] 2025-07-17T09:17:07.9567514Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:07.9568667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:07.9569761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 2%] 2025-07-17T09:17:07.9570871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 2%] 2025-07-17T09:17:07.9572100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 2%] 2025-07-17T09:17:07.9573251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:07.9574445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 2%] 2025-07-17T09:17:07.9575532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 2%] 2025-07-17T09:17:07.9576630Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:07.9577831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:07.9578927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 2%] 2025-07-17T09:17:07.9580032Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 2%] 2025-07-17T09:17:07.9581144Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0160s] [ 2%] 2025-07-17T09:17:07.9582230Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 2%] 2025-07-17T09:17:07.9583367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 2%] 2025-07-17T09:17:07.9584451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 2%] 2025-07-17T09:17:07.9585557Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 2%] 2025-07-17T09:17:07.9586680Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 2%] 2025-07-17T09:17:07.9587769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 2%] 2025-07-17T09:17:07.9589084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 2%] 2025-07-17T09:17:07.9590266Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0149s] [ 2%] 2025-07-17T09:17:07.9591446Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 2%] 2025-07-17T09:17:07.9592564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 2%] 2025-07-17T09:17:07.9593666Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 2%] 2025-07-17T09:17:07.9594851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 2%] 2025-07-17T09:17:07.9595960Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 2%] 2025-07-17T09:17:07.9597048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0041s] [ 2%] 2025-07-17T09:17:07.9598181Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 2%] 2025-07-17T09:17:07.9599292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:07.9600451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:07.9601554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 3%] 2025-07-17T09:17:07.9602661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 3%] 2025-07-17T09:17:07.9603777Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:07.9604892Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:07.9606201Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 3%] 2025-07-17T09:17:07.9607395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 3%] 2025-07-17T09:17:07.9608545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:07.9609666Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 3%] 2025-07-17T09:17:07.9610830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 3%] 2025-07-17T09:17:07.9611898Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 3%] 2025-07-17T09:17:07.9613036Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:07.9614120Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:07.9615209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 3%] 2025-07-17T09:17:07.9616320Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 3%] 2025-07-17T09:17:07.9617533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9618909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9620270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9621608Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9623164Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9624564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9625968Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9627306Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9628629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9629988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9631323Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9632689Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9634083Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9635419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9636913Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9638276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9639768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9641203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9642605Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9644006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9645333Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9646700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9648047Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9649430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9650921Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9652244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9653550Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9654935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9656392Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9657785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9659172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9660478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9661863Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9663243Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9664587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9665940Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9667288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9668675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9670004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9671338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9672806Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9674258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9675731Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9677043Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9678416Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9679750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9681085Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9682432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9683660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0067s] [ 3%] 2025-07-17T09:17:07.9684774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 3%] 2025-07-17T09:17:07.9685924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0050s] [ 3%] 2025-07-17T09:17:07.9687039Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0046s] [ 3%] 2025-07-17T09:17:07.9688229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 3%] 2025-07-17T09:17:07.9689470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 3%] 2025-07-17T09:17:07.9690712Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0050s] [ 3%] 2025-07-17T09:17:07.9691952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0045s] [ 3%] 2025-07-17T09:17:07.9693052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0047s] [ 3%] 2025-07-17T09:17:07.9694166Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 3%] 2025-07-17T09:17:07.9695287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0050s] [ 3%] 2025-07-17T09:17:07.9696376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0045s] [ 3%] 2025-07-17T09:17:07.9697507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0039s] [ 3%] 2025-07-17T09:17:07.9698605Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 3%] 2025-07-17T09:17:07.9699736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0048s] [ 3%] 2025-07-17T09:17:07.9700877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0044s] [ 3%] 2025-07-17T09:17:07.9701978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0052s] [ 3%] 2025-07-17T09:17:07.9703110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 3%] 2025-07-17T09:17:07.9704227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0054s] [ 3%] 2025-07-17T09:17:07.9705325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0049s] [ 3%] 2025-07-17T09:17:07.9706618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0043s] [ 3%] 2025-07-17T09:17:07.9707797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0043s] [ 3%] 2025-07-17T09:17:07.9708967Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0053s] [ 3%] 2025-07-17T09:17:07.9710135Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0049s] [ 3%] 2025-07-17T09:17:07.9711250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0051s] [ 3%] 2025-07-17T09:17:07.9712378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0042s] [ 3%] 2025-07-17T09:17:07.9713493Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0048s] [ 3%] 2025-07-17T09:17:07.9714642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0049s] [ 3%] 2025-07-17T09:17:07.9715765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0042s] [ 3%] 2025-07-17T09:17:07.9716876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0042s] [ 3%] 2025-07-17T09:17:07.9717983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0054s] [ 3%] 2025-07-17T09:17:07.9719106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0048s] [ 3%] 2025-07-17T09:17:07.9720205Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 3%] 2025-07-17T09:17:07.9721437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 3%] 2025-07-17T09:17:07.9722550Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 3%] 2025-07-17T09:17:07.9723794Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0045s] [ 3%] 2025-07-17T09:17:07.9724995Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 3%] 2025-07-17T09:17:07.9726159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 3%] 2025-07-17T09:17:07.9727309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 3%] 2025-07-17T09:17:07.9728415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0049s] [ 3%] 2025-07-17T09:17:07.9729515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0045s] [ 3%] 2025-07-17T09:17:07.9730645Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0039s] [ 3%] 2025-07-17T09:17:07.9731740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0048s] [ 3%] 2025-07-17T09:17:07.9732901Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0044s] [ 3%] 2025-07-17T09:17:07.9734008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0038s] [ 3%] 2025-07-17T09:17:07.9735112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 3%] 2025-07-17T09:17:07.9736231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0046s] [ 3%] 2025-07-17T09:17:07.9737368Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0044s] [ 3%] 2025-07-17T09:17:07.9738608Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9740054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9741517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9742933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9744280Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9745656Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9747067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9748429Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9749847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9751193Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:07.9752522Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9753986Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9755324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9756760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9758281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9759701Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9761100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9762460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9763811Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9765162Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9766518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9767889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9769238Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9770564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9772048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9773460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9774925Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9776299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9777617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9779026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9780369Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9781719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9783106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9784428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9785869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9787217Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9788586Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9790048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9791459Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9792867Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9794369Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9795719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9797084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9798513Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9799877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9801223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9802539Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9803915Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9805132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 4%] 2025-07-17T09:17:07.9806386Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:07.9807620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0047s] [ 4%] 2025-07-17T09:17:07.9808766Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 4%] 2025-07-17T09:17:07.9809939Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 4%] 2025-07-17T09:17:07.9811106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 4%] 2025-07-17T09:17:07.9812255Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0045s] [ 4%] 2025-07-17T09:17:07.9813379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 4%] 2025-07-17T09:17:07.9814487Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0045s] [ 4%] 2025-07-17T09:17:07.9815609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 4%] 2025-07-17T09:17:07.9816707Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 4%] 2025-07-17T09:17:07.9817790Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 4%] 2025-07-17T09:17:07.9818930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 4%] 2025-07-17T09:17:07.9820048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 4%] 2025-07-17T09:17:07.9821201Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0046s] [ 4%] 2025-07-17T09:17:07.9822334Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 4%] 2025-07-17T09:17:07.9823521Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0059s] [ 4%] 2025-07-17T09:17:07.9824767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:07.9825930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 4%] 2025-07-17T09:17:07.9827036Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 4%] 2025-07-17T09:17:07.9828178Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:07.9829291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 4%] 2025-07-17T09:17:07.9830428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 4%] 2025-07-17T09:17:07.9831526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 4%] 2025-07-17T09:17:07.9832618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0059s] [ 4%] 2025-07-17T09:17:07.9833818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:07.9834999Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0048s] [ 4%] 2025-07-17T09:17:07.9836164Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0041s] [ 4%] 2025-07-17T09:17:07.9837257Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 4%] 2025-07-17T09:17:07.9838334Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:07.9839477Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0050s] [ 4%] 2025-07-17T09:17:07.9840725Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0041s] [ 4%] 2025-07-17T09:17:07.9841910Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:07.9843088Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 4%] 2025-07-17T09:17:07.9844258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0048s] [ 4%] 2025-07-17T09:17:07.9845384Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 4%] 2025-07-17T09:17:07.9846499Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:07.9847593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:07.9848739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0048s] [ 4%] 2025-07-17T09:17:07.9849862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 4%] 2025-07-17T09:17:07.9851019Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 4%] 2025-07-17T09:17:07.9852122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:07.9882307Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0048s] [ 4%] 2025-07-17T09:17:07.9883445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0039s] [ 4%] 2025-07-17T09:17:07.9884456Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 4%] 2025-07-17T09:17:07.9885445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 4%] 2025-07-17T09:17:07.9886689Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0047s] [ 4%] 2025-07-17T09:17:07.9887773Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0040s] [ 4%] 2025-07-17T09:17:07.9888964Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9890204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9891434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9892654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9893886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9895122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9896347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9897569Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9898790Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9900016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9901367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9902647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9903933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9905159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9906378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9907595Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9908824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9910052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9911276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9912504Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9913733Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9915055Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9916429Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9917768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9919209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9920458Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9921695Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9922914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9924141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9925363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9926581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9927792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9929017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9930245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9931594Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9932877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9934168Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:07.9935402Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:07.9936625Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:07.9937851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:07.9939073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:07.9940289Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:07.9941511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:07.9942727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:07.9943943Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:07.9945183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:07.9946537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:07.9947813Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:07.9948984Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0064s] [ 5%] 2025-07-17T09:17:07.9949982Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 5%] 2025-07-17T09:17:07.9950969Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 5%] 2025-07-17T09:17:07.9951958Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 5%] 2025-07-17T09:17:07.9952944Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0028s] [ 5%] 2025-07-17T09:17:07.9954009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 5%] 2025-07-17T09:17:07.9955004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 5%] 2025-07-17T09:17:07.9955992Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 5%] 2025-07-17T09:17:07.9956969Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0048s] [ 5%] 2025-07-17T09:17:07.9957950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 5%] 2025-07-17T09:17:07.9958931Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 5%] 2025-07-17T09:17:07.9959899Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 5%] 2025-07-17T09:17:07.9960869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0028s] [ 5%] 2025-07-17T09:17:07.9962078Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 5%] 2025-07-17T09:17:07.9963113Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 5%] 2025-07-17T09:17:07.9964083Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 5%] 2025-07-17T09:17:07.9965062Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0050s] [ 5%] 2025-07-17T09:17:07.9966057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 5%] 2025-07-17T09:17:07.9967045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 5%] 2025-07-17T09:17:07.9968015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 5%] 2025-07-17T09:17:07.9969004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 5%] 2025-07-17T09:17:07.9969997Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 5%] 2025-07-17T09:17:07.9970982Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 5%] 2025-07-17T09:17:07.9971962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 5%] 2025-07-17T09:17:07.9972991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 ('RERUN', {'yellow': True}) [0.0052s] [ 5%] 2025-07-17T09:17:07.9974051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 ('RERUN', {'yellow': True}) [0.0031s] [ 5%] 2025-07-17T09:17:07.9975066Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 FAILED [0.0030s] [ 5%] 2025-07-17T09:17:07.9975678Z 2025-07-17T09:17:07.9975771Z ==================================== RERUNS ==================================== 2025-07-17T09:17:07.9976359Z _ TestSDPACudaOnlyCUDA.test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 _ 2025-07-17T09:17:07.9976892Z Traceback (most recent call last): 2025-07-17T09:17:07.9977312Z File "/var/lib/jenkins/pytorch/test/test_transformers.py", line 3657, in test_flash_attention_vs_math_ref_grads 2025-07-17T09:17:07.9977730Z check_out_and_grad( 2025-07-17T09:17:07.9978019Z File "/var/lib/jenkins/pytorch/test/test_transformers.py", line 182, in check_out_and_grad 2025-07-17T09:17:07.9978453Z _check_equal(ref_grad, lp_ref_grad, comp_grad, fudge_factors.get(name, default_fudge_factor), name) 2025-07-17T09:17:07.9978868Z File "/var/lib/jenkins/pytorch/test/test_transformers.py", line 142, in _check_equal 2025-07-17T09:17:07.9979165Z raise ValueError(msg) 2025-07-17T09:17:07.9979478Z ValueError: grad_value Test error 6.251037120819092e-05 is greater than threshold 6.103515625e-05! 2025-07-17T09:17:07.9979731Z 2025-07-17T09:17:07.9979858Z To execute this test, run the following from the base repo dir: 2025-07-17T09:17:07.9980567Z PYTORCH_TEST_WITH_ROCM=1 python test/test_transformers.py TestSDPACudaOnlyCUDA.test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 2025-07-17T09:17:07.9981145Z 2025-07-17T09:17:07.9981295Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-07-17T09:17:07.9981889Z _ TestSDPACudaOnlyCUDA.test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 _ 2025-07-17T09:17:07.9982410Z Traceback (most recent call last): 2025-07-17T09:17:07.9982766Z File "/var/lib/jenkins/pytorch/test/test_transformers.py", line 3657, in test_flash_attention_vs_math_ref_grads 2025-07-17T09:17:07.9983121Z check_out_and_grad( 2025-07-17T09:17:07.9983408Z File "/var/lib/jenkins/pytorch/test/test_transformers.py", line 182, in check_out_and_grad 2025-07-17T09:17:07.9983832Z _check_equal(ref_grad, lp_ref_grad, comp_grad, fudge_factors.get(name, default_fudge_factor), name) 2025-07-17T09:17:07.9984247Z File "/var/lib/jenkins/pytorch/test/test_transformers.py", line 142, in _check_equal 2025-07-17T09:17:07.9984548Z raise ValueError(msg) 2025-07-17T09:17:07.9984845Z ValueError: grad_value Test error 6.251037120819092e-05 is greater than threshold 6.103515625e-05! 2025-07-17T09:17:07.9985088Z 2025-07-17T09:17:07.9985211Z To execute this test, run the following from the base repo dir: 2025-07-17T09:17:07.9985902Z PYTORCH_TEST_WITH_ROCM=1 python test/test_transformers.py TestSDPACudaOnlyCUDA.test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 2025-07-17T09:17:07.9986465Z 2025-07-17T09:17:07.9986615Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-07-17T09:17:07.9986903Z =================================== FAILURES =================================== 2025-07-17T09:17:07.9987434Z _ TestSDPACudaOnlyCUDA.test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 _ 2025-07-17T09:17:07.9987948Z Traceback (most recent call last): 2025-07-17T09:17:07.9988287Z File "/var/lib/jenkins/pytorch/test/test_transformers.py", line 3657, in test_flash_attention_vs_math_ref_grads 2025-07-17T09:17:07.9988637Z check_out_and_grad( 2025-07-17T09:17:07.9988910Z File "/var/lib/jenkins/pytorch/test/test_transformers.py", line 182, in check_out_and_grad 2025-07-17T09:17:07.9989330Z _check_equal(ref_grad, lp_ref_grad, comp_grad, fudge_factors.get(name, default_fudge_factor), name) 2025-07-17T09:17:07.9989817Z File "/var/lib/jenkins/pytorch/test/test_transformers.py", line 142, in _check_equal 2025-07-17T09:17:07.9990168Z raise ValueError(msg) 2025-07-17T09:17:07.9990463Z ValueError: grad_value Test error 6.251037120819092e-05 is greater than threshold 6.103515625e-05! 2025-07-17T09:17:07.9990713Z 2025-07-17T09:17:07.9990881Z To execute this test, run the following from the base repo dir: 2025-07-17T09:17:07.9991612Z PYTORCH_TEST_WITH_ROCM=1 python test/test_transformers.py TestSDPACudaOnlyCUDA.test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 2025-07-17T09:17:07.9992191Z 2025-07-17T09:17:07.9992338Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-07-17T09:17:07.9992872Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/test_transformers/test_transformers-c3d3be2371ce67e4.xml - 2025-07-17T09:17:07.9993347Z =========================== short test summary info ============================ 2025-07-17T09:17:07.9994276Z FAILED [0.0030s] test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 - ValueError: grad_value Test error 6.251037120819092e-05 is greater than threshold 6.103515625e-05! 2025-07-17T09:17:07.9995002Z 2025-07-17T09:17:07.9995119Z To execute this test, run the following from the base repo dir: 2025-07-17T09:17:07.9995808Z PYTORCH_TEST_WITH_ROCM=1 python test/test_transformers.py TestSDPACudaOnlyCUDA.test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 2025-07-17T09:17:07.9996376Z 2025-07-17T09:17:07.9996518Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-07-17T09:17:07.9996819Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-07-17T09:17:07.9997090Z ============= 1 failed, 355 passed, 292 skipped, 2 rerun in 20.75s ============= 2025-07-17T09:17:07.9997327Z Got exit code 1 2025-07-17T09:17:07.9997484Z Retrying single test... 2025-07-17T09:17:07.9998325Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T09:17:07.9999138Z import pkg_resources 2025-07-17T09:17:07.9999497Z Test results will be stored in test-reports/python-pytest/test_transformers/test_transformers-6ed5a93db89cf6a3.xml 2025-07-17T09:17:07.9999895Z ============================= test session starts ============================== 2025-07-17T09:17:08.0000242Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T09:17:08.0000550Z cachedir: .pytest_cache 2025-07-17T09:17:08.0000915Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T09:17:08.0001308Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T09:17:08.0001496Z configfile: pytest.ini 2025-07-17T09:17:08.0001874Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T09:17:08.0002321Z collecting ... collected 12242 items / 12241 deselected / 1 selected 2025-07-17T09:17:08.0003089Z stepcurrent: skipping 647 already run items. Running only test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 2025-07-17T09:17:08.0003875Z Running 1 items in this shard 2025-07-17T09:17:08.0003998Z 2025-07-17T09:17:08.0004807Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 [W717 09:10:10.329553327 attention.cpp:916] Warning: Dropout mask should only be used for testing purposes. (function operator()) 2025-07-17T09:17:08.0005688Z PASSED [1.4877s] [100%] 2025-07-17T09:17:08.0005796Z 2025-07-17T09:17:08.0006115Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/test_transformers/test_transformers-6ed5a93db89cf6a3.xml - 2025-07-17T09:17:08.0006592Z ===================== 1 passed, 12241 deselected in 2.07s ====================== 2025-07-17T09:17:08.0006817Z Got exit code 0 2025-07-17T09:17:08.0007045Z Test succeeeded in new process, continuing with the rest of the tests 2025-07-17T09:17:08.0007968Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T09:17:08.0008786Z import pkg_resources 2025-07-17T09:17:08.0009133Z Test results will be stored in test-reports/python-pytest/test_transformers/test_transformers-929fa49726c73d6e.xml 2025-07-17T09:17:08.0009525Z ============================= test session starts ============================== 2025-07-17T09:17:08.0009863Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T09:17:08.0010167Z cachedir: .pytest_cache 2025-07-17T09:17:08.0010530Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T09:17:08.0010925Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T09:17:08.0011116Z configfile: pytest.ini 2025-07-17T09:17:08.0011498Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T09:17:08.0011955Z collecting ... collected 12242 items / 648 deselected / 11594 selected 2025-07-17T09:17:08.0012229Z stepcurrent: skipping 648 already run items. 2025-07-17T09:17:08.0012438Z Running 11594 items in this shard 2025-07-17T09:17:08.0012563Z 2025-07-17T09:17:08.0013261Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 [W717 09:10:19.979447697 attention.cpp:916] Warning: Dropout mask should only be used for testing purposes. (function operator()) 2025-07-17T09:17:08.0014070Z PASSED [1.5368s] [ 0%] 2025-07-17T09:17:08.0014637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0343s] [ 0%] 2025-07-17T09:17:08.0015639Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0183s] [ 0%] 2025-07-17T09:17:08.0016633Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0053s] [ 0%] 2025-07-17T09:17:08.0017624Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0044s] [ 0%] 2025-07-17T09:17:08.0018682Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0048s] [ 0%] 2025-07-17T09:17:08.0019801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0047s] [ 0%] 2025-07-17T09:17:08.0020866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0157s] [ 0%] 2025-07-17T09:17:08.0021871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0045s] [ 0%] 2025-07-17T09:17:08.0022866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0052s] [ 0%] 2025-07-17T09:17:08.0023849Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0047s] [ 0%] 2025-07-17T09:17:08.0024844Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0043s] [ 0%] 2025-07-17T09:17:08.0025839Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 0%] 2025-07-17T09:17:08.0026837Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0048s] [ 0%] 2025-07-17T09:17:08.0027835Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0047s] [ 0%] 2025-07-17T09:17:08.0028826Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0043s] [ 0%] 2025-07-17T09:17:08.0029804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0042s] [ 0%] 2025-07-17T09:17:08.0030786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0047s] [ 0%] 2025-07-17T09:17:08.0031768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0047s] [ 0%] 2025-07-17T09:17:08.0032761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0043s] [ 0%] 2025-07-17T09:17:08.0033930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0043s] [ 0%] 2025-07-17T09:17:08.0035003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0047s] [ 0%] 2025-07-17T09:17:08.0036060Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0047s] [ 0%] 2025-07-17T09:17:08.0037190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0038441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0039676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0040903Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0042137Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0043370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0044609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0045855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0047101Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0048341Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0049789Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0051072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0052294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0053528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0054763Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0055976Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0057223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0058475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0059710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0060938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0062183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0063422Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0064838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0066123Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0067350Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0068577Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0069801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0071016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0072240Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0073463Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0074755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0075984Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0077218Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0078453Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0079871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0081171Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0082407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0083645Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0084876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0086106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0087337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0088565Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0089785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0091007Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0092227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0093461Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0094864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0096135Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 0%] 2025-07-17T09:17:08.0097246Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0152s] [ 0%] 2025-07-17T09:17:08.0098251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 0%] 2025-07-17T09:17:08.0099250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 0%] 2025-07-17T09:17:08.0100237Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 0%] 2025-07-17T09:17:08.0101226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 0%] 2025-07-17T09:17:08.0102219Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 0%] 2025-07-17T09:17:08.0103206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 0%] 2025-07-17T09:17:08.0104192Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 0%] 2025-07-17T09:17:08.0105171Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0133s] [ 0%] 2025-07-17T09:17:08.0106159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 0%] 2025-07-17T09:17:08.0107135Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 0%] 2025-07-17T09:17:08.0108116Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 0%] 2025-07-17T09:17:08.0109152Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 0%] 2025-07-17T09:17:08.0110247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 0%] 2025-07-17T09:17:08.0111278Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 0%] 2025-07-17T09:17:08.0112258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 0%] 2025-07-17T09:17:08.0113244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0160s] [ 0%] 2025-07-17T09:17:08.0114303Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0045s] [ 0%] 2025-07-17T09:17:08.0115296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0048s] [ 0%] 2025-07-17T09:17:08.0116279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0047s] [ 0%] 2025-07-17T09:17:08.0117275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0045s] [ 0%] 2025-07-17T09:17:08.0118288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0044s] [ 0%] 2025-07-17T09:17:08.0119290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 0%] 2025-07-17T09:17:08.0120284Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0047s] [ 0%] 2025-07-17T09:17:08.0121270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0156s] [ 0%] 2025-07-17T09:17:08.0122250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 0%] 2025-07-17T09:17:08.0123224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 0%] 2025-07-17T09:17:08.0124354Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 0%] 2025-07-17T09:17:08.0125403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 0%] 2025-07-17T09:17:08.0126449Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 0%] 2025-07-17T09:17:08.0127428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 0%] 2025-07-17T09:17:08.0128408Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 0%] 2025-07-17T09:17:08.0129392Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 0%] 2025-07-17T09:17:08.0130387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 0%] 2025-07-17T09:17:08.0131371Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 0%] 2025-07-17T09:17:08.0132359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 0%] 2025-07-17T09:17:08.0133345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 0%] 2025-07-17T09:17:08.0134344Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 0%] 2025-07-17T09:17:08.0135342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 0%] 2025-07-17T09:17:08.0136328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 0%] 2025-07-17T09:17:08.0137312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 0%] 2025-07-17T09:17:08.0138289Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 0%] 2025-07-17T09:17:08.0139385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 0%] 2025-07-17T09:17:08.0140414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 0%] 2025-07-17T09:17:08.0141457Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 1%] 2025-07-17T09:17:08.0142445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 1%] 2025-07-17T09:17:08.0143427Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 1%] 2025-07-17T09:17:08.0144402Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 1%] 2025-07-17T09:17:08.0145516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0146751Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0147992Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0149213Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0150434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0151659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0152882Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0154296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0155647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0156933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0158156Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0159372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0160580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0161795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0163013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0164224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0165454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0166683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0167903Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0169183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0170515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0171814Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0173051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0174279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0175507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0176722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0177949Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0179167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0180385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0181604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0182816Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0184106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0185439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0186725Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0187961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0189188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0190420Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0191649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0192879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0194173Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0195408Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0196626Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0197841Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0199123Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0200481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0201768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0202991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0204202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0205309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0137s] [ 1%] 2025-07-17T09:17:08.0206302Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 1%] 2025-07-17T09:17:08.0207302Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:08.0208293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:08.0209283Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0210276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0211271Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 1%] 2025-07-17T09:17:08.0212262Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0213247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0122s] [ 1%] 2025-07-17T09:17:08.0214305Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0215397Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:08.0216433Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:08.0217414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 1%] 2025-07-17T09:17:08.0218404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 1%] 2025-07-17T09:17:08.0219388Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 1%] 2025-07-17T09:17:08.0220367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 1%] 2025-07-17T09:17:08.0221351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0125s] [ 1%] 2025-07-17T09:17:08.0222352Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 1%] 2025-07-17T09:17:08.0223338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 1%] 2025-07-17T09:17:08.0224329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:08.0225332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0226342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0227342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:08.0228332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 1%] 2025-07-17T09:17:08.0229379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0113s] [ 1%] 2025-07-17T09:17:08.0230468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 1%] 2025-07-17T09:17:08.0231542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:08.0232526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 1%] 2025-07-17T09:17:08.0233518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0234573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0235558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:08.0236544Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 1%] 2025-07-17T09:17:08.0237527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0238518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 1%] 2025-07-17T09:17:08.0239504Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:08.0240486Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 1%] 2025-07-17T09:17:08.0241479Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0242475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0243462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:08.0244528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 1%] 2025-07-17T09:17:08.0245634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 1%] 2025-07-17T09:17:08.0246683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0247680Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 1%] 2025-07-17T09:17:08.0248667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 1%] 2025-07-17T09:17:08.0249654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0250644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 1%] 2025-07-17T09:17:08.0251626Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 1%] 2025-07-17T09:17:08.0252611Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 1%] 2025-07-17T09:17:08.0253738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0254982Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0256215Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0257438Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0258698Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0260071Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0261379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0262690Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0263921Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0265150Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0266367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0267576Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0268795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0270015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0271231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0272439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 1%] 2025-07-17T09:17:08.0273661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0275109Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0276419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0277710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0278942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0280181Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0281416Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0282646Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0283872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0285103Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0286318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0287537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0288754Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0290070Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0291438Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0292709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0293935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0295166Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0296405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0297627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0298860Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0300095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0301327Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0302562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0303793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0305110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0306439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0307737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0308967Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0310196Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0311418Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0312634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0313750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0127s] [ 2%] 2025-07-17T09:17:08.0314826Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 2%] 2025-07-17T09:17:08.0315810Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 2%] 2025-07-17T09:17:08.0316798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:08.0317785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0027s] [ 2%] 2025-07-17T09:17:08.0318777Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 2%] 2025-07-17T09:17:08.0319758Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:08.0320889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:08.0321942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0110s] [ 2%] 2025-07-17T09:17:08.0322981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 2%] 2025-07-17T09:17:08.0323956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0029s] [ 2%] 2025-07-17T09:17:08.0324937Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0029s] [ 2%] 2025-07-17T09:17:08.0325909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0027s] [ 2%] 2025-07-17T09:17:08.0326882Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 2%] 2025-07-17T09:17:08.0327852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0029s] [ 2%] 2025-07-17T09:17:08.0328822Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0029s] [ 2%] 2025-07-17T09:17:08.0329801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0048s] [ 2%] 2025-07-17T09:17:08.0330786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 2%] 2025-07-17T09:17:08.0331784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:08.0332771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 2%] 2025-07-17T09:17:08.0333777Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 2%] 2025-07-17T09:17:08.0334764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 2%] 2025-07-17T09:17:08.0335874Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:08.0336920Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 2%] 2025-07-17T09:17:08.0337967Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0045s] [ 2%] 2025-07-17T09:17:08.0338961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:08.0339941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:08.0340917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:08.0341892Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:08.0342349Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:08.0342813Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:08.0343267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:08.0343736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:08.0344206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 2%] 2025-07-17T09:17:08.0344670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:08.0345136Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:08.0345603Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 2%] 2025-07-17T09:17:08.0346227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 2%] 2025-07-17T09:17:08.0346741Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:08.0347259Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:08.0347722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:08.0348184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:08.0348647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:08.0349100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 2%] 2025-07-17T09:17:08.0349569Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 2%] 2025-07-17T09:17:08.0350030Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 2%] 2025-07-17T09:17:08.0350492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 2%] 2025-07-17T09:17:08.0350952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 2%] 2025-07-17T09:17:08.0351555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0352154Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0352734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0353432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0354155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0354800Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0355393Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0355970Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0356552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0357132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0357705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0358290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0358865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0359451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0360038Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0360669Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0361378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0362016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0362605Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0363189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0363776Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0364368Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0364954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0365543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0366131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0366712Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0367294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0367944Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0368644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0369277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0369856Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0370450Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0371035Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0371630Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0372217Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0372795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 2%] 2025-07-17T09:17:08.0373390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0373975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0374560Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0375211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0375887Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0376518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0377095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0377682Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0378266Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0378842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0379435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0380011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0380492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0069s] [ 3%] 2025-07-17T09:17:08.0380966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 3%] 2025-07-17T09:17:08.0381434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:08.0381904Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 3%] 2025-07-17T09:17:08.0382376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:08.0383007Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:08.0383528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 3%] 2025-07-17T09:17:08.0384001Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 3%] 2025-07-17T09:17:08.0384475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0058s] [ 3%] 2025-07-17T09:17:08.0384938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:08.0385410Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:08.0385879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 3%] 2025-07-17T09:17:08.0386362Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:08.0386834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:08.0387306Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:08.0387780Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 3%] 2025-07-17T09:17:08.0388253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 3%] 2025-07-17T09:17:08.0388727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 3%] 2025-07-17T09:17:08.0389191Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 3%] 2025-07-17T09:17:08.0389742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 3%] 2025-07-17T09:17:08.0390316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 3%] 2025-07-17T09:17:08.0390847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 3%] 2025-07-17T09:17:08.0391321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 3%] 2025-07-17T09:17:08.0391792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 3%] 2025-07-17T09:17:08.0392262Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0060s] [ 3%] 2025-07-17T09:17:08.0392728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 3%] 2025-07-17T09:17:08.0393197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 3%] 2025-07-17T09:17:08.0393662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 3%] 2025-07-17T09:17:08.0394241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 3%] 2025-07-17T09:17:08.0394714Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 3%] 2025-07-17T09:17:08.0395177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 3%] 2025-07-17T09:17:08.0395648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 3%] 2025-07-17T09:17:08.0396119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 3%] 2025-07-17T09:17:08.0396590Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 3%] 2025-07-17T09:17:08.0397194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 3%] 2025-07-17T09:17:08.0397735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 3%] 2025-07-17T09:17:08.0398286Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 3%] 2025-07-17T09:17:08.0398755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 3%] 2025-07-17T09:17:08.0399232Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 3%] 2025-07-17T09:17:08.0399707Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 3%] 2025-07-17T09:17:08.0400183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 3%] 2025-07-17T09:17:08.0400653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 3%] 2025-07-17T09:17:08.0401123Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 3%] 2025-07-17T09:17:08.0401591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 3%] 2025-07-17T09:17:08.0402056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 3%] 2025-07-17T09:17:08.0402526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 3%] 2025-07-17T09:17:08.0402990Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 3%] 2025-07-17T09:17:08.0403458Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 3%] 2025-07-17T09:17:08.0404052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0404807Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0405455Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0406040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0406643Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0407233Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0407823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0408419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0409004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0409591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0410174Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0410757Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0411346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0412080Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0412713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0413296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0413899Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0414498Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0415086Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0415683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0416279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0416871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0417471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0418059Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0418662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0419414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0420054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0420644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0421235Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0421825Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0422414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0422993Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0423590Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0424173Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0424767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0425358Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0425944Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0426705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0427391Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0427982Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0428576Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0429162Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0429747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0430328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0430927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0431516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0432101Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0432690Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 3%] 2025-07-17T09:17:08.0433162Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0068s] [ 3%] 2025-07-17T09:17:08.0433776Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 3%] 2025-07-17T09:17:08.0434403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 3%] 2025-07-17T09:17:08.0434943Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 3%] 2025-07-17T09:17:08.0435435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 3%] 2025-07-17T09:17:08.0435915Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 3%] 2025-07-17T09:17:08.0436393Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 3%] 2025-07-17T09:17:08.0436861Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 3%] 2025-07-17T09:17:08.0437334Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0055s] [ 4%] 2025-07-17T09:17:08.0437805Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 4%] 2025-07-17T09:17:08.0438268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 4%] 2025-07-17T09:17:08.0438730Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 4%] 2025-07-17T09:17:08.0439199Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 4%] 2025-07-17T09:17:08.0439668Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 4%] 2025-07-17T09:17:08.0440127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 4%] 2025-07-17T09:17:08.0440594Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 4%] 2025-07-17T09:17:08.0441204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0061s] [ 4%] 2025-07-17T09:17:08.0441721Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 4%] 2025-07-17T09:17:08.0442241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 4%] 2025-07-17T09:17:08.0442710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:08.0443199Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:08.0443672Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:08.0444147Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 4%] 2025-07-17T09:17:08.0444619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:08.0445091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0058s] [ 4%] 2025-07-17T09:17:08.0445561Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:08.0446021Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 4%] 2025-07-17T09:17:08.0446497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:08.0446966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:08.0447440Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:08.0447909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 4%] 2025-07-17T09:17:08.0448476Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:08.0449001Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:08.0449526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:08.0450003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 4%] 2025-07-17T09:17:08.0450478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 4%] 2025-07-17T09:17:08.0450957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:08.0451428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:08.0451898Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 4%] 2025-07-17T09:17:08.0452367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:08.0452830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:08.0453297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:08.0453759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 4%] 2025-07-17T09:17:08.0454229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:08.0454702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:08.0455164Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:08.0455802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 4%] 2025-07-17T09:17:08.0456315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:08.0456924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0457517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0458121Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0458722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0459317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0459915Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0460502Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0461092Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0461682Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0462279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0462969Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0463595Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0464236Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0464826Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0465404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0465997Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0466588Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0467182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0467774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0468362Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0468958Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0469554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0470255Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0470896Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0471534Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0472123Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0472709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0473288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0473952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0474538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0475131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0475719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0476313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0476907Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0477622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0478290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0478944Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0479540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0480134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0480723Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0481319Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0481911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0482493Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0483083Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0483668Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0484261Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0484978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0485612Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 4%] 2025-07-17T09:17:08.0486136Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0058s] [ 4%] 2025-07-17T09:17:08.0486605Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0027s] [ 4%] 2025-07-17T09:17:08.0487079Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 4%] 2025-07-17T09:17:08.0487539Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 4%] 2025-07-17T09:17:08.0488013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0027s] [ 4%] 2025-07-17T09:17:08.0488487Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 4%] 2025-07-17T09:17:08.0488954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 4%] 2025-07-17T09:17:08.0489426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 4%] 2025-07-17T09:17:08.0489886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0047s] [ 4%] 2025-07-17T09:17:08.0490359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 4%] 2025-07-17T09:17:08.0490820Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0029s] [ 4%] 2025-07-17T09:17:08.0491280Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 4%] 2025-07-17T09:17:08.0491749Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0027s] [ 4%] 2025-07-17T09:17:08.0492379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 4%] 2025-07-17T09:17:08.0492905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0029s] [ 4%] 2025-07-17T09:17:08.0493364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0029s] [ 4%] 2025-07-17T09:17:08.0493841Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 4%] 2025-07-17T09:17:08.0494311Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 4%] 2025-07-17T09:17:08.0494779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 4%] 2025-07-17T09:17:08.0495247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0077s] [ 4%] 2025-07-17T09:17:08.0495718Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0112s] [ 4%] 2025-07-17T09:17:08.0496193Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 4%] 2025-07-17T09:17:08.0496662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 4%] 2025-07-17T09:17:08.0497134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 4%] 2025-07-17T09:17:08.0497597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 4%] 2025-07-17T09:17:08.0498062Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 4%] 2025-07-17T09:17:08.0498528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 4%] 2025-07-17T09:17:08.0499049Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 4%] 2025-07-17T09:17:08.0499619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 5%] 2025-07-17T09:17:08.0500133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 5%] 2025-07-17T09:17:08.0500603Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 5%] 2025-07-17T09:17:08.0501069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 5%] 2025-07-17T09:17:08.0501542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 5%] 2025-07-17T09:17:08.0502010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0167s] [ 5%] 2025-07-17T09:17:08.0502469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 5%] 2025-07-17T09:17:08.0502941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 5%] 2025-07-17T09:17:08.0503407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 5%] 2025-07-17T09:17:08.0503882Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 5%] 2025-07-17T09:17:08.0504349Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 5%] 2025-07-17T09:17:08.0504823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 5%] 2025-07-17T09:17:08.0505287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 5%] 2025-07-17T09:17:08.0505749Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 5%] 2025-07-17T09:17:08.0506266Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 5%] 2025-07-17T09:17:08.0506831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 5%] 2025-07-17T09:17:08.0507366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 5%] 2025-07-17T09:17:08.0507824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 5%] 2025-07-17T09:17:08.0508296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 5%] 2025-07-17T09:17:08.0508761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 5%] 2025-07-17T09:17:08.0509355Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0509945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0510531Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0511124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0511719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0512316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0512903Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0513536Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0514304Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0514953Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0515542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0516131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0516711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0517298Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0517880Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0518454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0519041Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0519628Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0520214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0520797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0521556Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0522216Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0522805Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0523396Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0523991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0524571Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0525161Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0525738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0526329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0526916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0527496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0528074Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0528828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0529469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0530060Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0530651Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0531242Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0531831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0532431Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0533020Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0533602Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0534192Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0534767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0535353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0536115Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0536761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0537348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0537934Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 5%] 2025-07-17T09:17:08.0538420Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0054s] [ 5%] 2025-07-17T09:17:08.0538897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 5%] 2025-07-17T09:17:08.0539370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 5%] 2025-07-17T09:17:08.0539845Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0055s] [ 5%] 2025-07-17T09:17:08.0540318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 5%] 2025-07-17T09:17:08.0540794Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 5%] 2025-07-17T09:17:08.0541270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0045s] [ 5%] 2025-07-17T09:17:08.0541740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0055s] [ 5%] 2025-07-17T09:17:08.0542213Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0042s] [ 5%] 2025-07-17T09:17:08.0542673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 5%] 2025-07-17T09:17:08.0544141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 5%] 2025-07-17T09:17:08.0544659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0054s] [ 5%] 2025-07-17T09:17:08.0545142Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 5%] 2025-07-17T09:17:08.0545618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 5%] 2025-07-17T09:17:08.0546086Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0044s] [ 5%] 2025-07-17T09:17:08.0546555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0054s] [ 5%] 2025-07-17T09:17:08.0547029Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 5%] 2025-07-17T09:17:08.0547510Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 5%] 2025-07-17T09:17:08.0547979Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 5%] 2025-07-17T09:17:08.0548454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0060s] [ 5%] 2025-07-17T09:17:08.0548933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 5%] 2025-07-17T09:17:08.0549407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 5%] 2025-07-17T09:17:08.0549886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 5%] 2025-07-17T09:17:08.0550356Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0059s] [ 5%] 2025-07-17T09:17:08.0550894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 5%] 2025-07-17T09:17:08.0551463Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 5%] 2025-07-17T09:17:08.0551996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0049s] [ 5%] 2025-07-17T09:17:08.0552465Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0061s] [ 5%] 2025-07-17T09:17:08.0552937Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 5%] 2025-07-17T09:17:08.0553413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 5%] 2025-07-17T09:17:08.0553954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0049s] [ 5%] 2025-07-17T09:17:08.0554425Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0059s] [ 5%] 2025-07-17T09:17:08.0554901Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 5%] 2025-07-17T09:17:08.0555376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 5%] 2025-07-17T09:17:08.0555851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0045s] [ 5%] 2025-07-17T09:17:08.0556321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0055s] [ 5%] 2025-07-17T09:17:08.0556801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 5%] 2025-07-17T09:17:08.0557273Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 5%] 2025-07-17T09:17:08.0557749Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0045s] [ 5%] 2025-07-17T09:17:08.0558348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0056s] [ 5%] 2025-07-17T09:17:08.0558884Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 5%] 2025-07-17T09:17:08.0559413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 5%] 2025-07-17T09:17:08.0559880Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 5%] 2025-07-17T09:17:08.0560473Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0056s] [ 5%] 2025-07-17T09:17:08.0561100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 5%] 2025-07-17T09:17:08.0561622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 5%] 2025-07-17T09:17:08.0562165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 5%] 2025-07-17T09:17:08.0562674Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0055s] [ 5%] 2025-07-17T09:17:08.0563374Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0564070Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0579602Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0580224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0580838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0581734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0582388Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0582985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0583582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0584163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0584746Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0585330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0585914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0586509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0587104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0587691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0588285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0589051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0589698Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0590281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0590888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0591479Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0592079Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0592674Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0593265Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0593851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0594516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0595107Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0595693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0596473Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0597150Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0597732Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0598334Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0598938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0599522Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0600114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0600706Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0601297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0601895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0602483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0603071Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0603853Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0604486Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0605067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0605654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0606245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0606828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0607415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0607899Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 6%] 2025-07-17T09:17:08.0608367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 6%] 2025-07-17T09:17:08.0608846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0044s] [ 6%] 2025-07-17T09:17:08.0609309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 6%] 2025-07-17T09:17:08.0609795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 6%] 2025-07-17T09:17:08.0610270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 6%] 2025-07-17T09:17:08.0610863Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 6%] 2025-07-17T09:17:08.0611387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 6%] 2025-07-17T09:17:08.0611906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 6%] 2025-07-17T09:17:08.0612378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 6%] 2025-07-17T09:17:08.0612847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0041s] [ 6%] 2025-07-17T09:17:08.0613326Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0040s] [ 6%] 2025-07-17T09:17:08.0613804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 6%] 2025-07-17T09:17:08.0614271Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 6%] 2025-07-17T09:17:08.0614743Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 6%] 2025-07-17T09:17:08.0615202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0040s] [ 6%] 2025-07-17T09:17:08.0615679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 6%] 2025-07-17T09:17:08.0616149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 6%] 2025-07-17T09:17:08.0616622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0043s] [ 6%] 2025-07-17T09:17:08.0617094Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 6%] 2025-07-17T09:17:08.0617572Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 6%] 2025-07-17T09:17:08.0618204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 6%] 2025-07-17T09:17:08.0618730Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0044s] [ 6%] 2025-07-17T09:17:08.0619204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 6%] 2025-07-17T09:17:08.0619670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 6%] 2025-07-17T09:17:08.0620144Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 6%] 2025-07-17T09:17:08.0620609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0042s] [ 6%] 2025-07-17T09:17:08.0621068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0041s] [ 6%] 2025-07-17T09:17:08.0621545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 6%] 2025-07-17T09:17:08.0622012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 6%] 2025-07-17T09:17:08.0622487Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0042s] [ 6%] 2025-07-17T09:17:08.0622948Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0041s] [ 6%] 2025-07-17T09:17:08.0623424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 6%] 2025-07-17T09:17:08.0623894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 6%] 2025-07-17T09:17:08.0624359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 6%] 2025-07-17T09:17:08.0624881Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 6%] 2025-07-17T09:17:08.0625478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 6%] 2025-07-17T09:17:08.0626007Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 6%] 2025-07-17T09:17:08.0626476Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 6%] 2025-07-17T09:17:08.0626950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 6%] 2025-07-17T09:17:08.0627424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 6%] 2025-07-17T09:17:08.0627890Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 6%] 2025-07-17T09:17:08.0628357Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 6%] 2025-07-17T09:17:08.0628820Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0041s] [ 6%] 2025-07-17T09:17:08.0629294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 6%] 2025-07-17T09:17:08.0629759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 6%] 2025-07-17T09:17:08.0630230Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 6%] 2025-07-17T09:17:08.0630696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0041s] [ 6%] 2025-07-17T09:17:08.0631297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0631888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0632631Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0633272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0633866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0634517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0635110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0635700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0636295Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0636887Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0637468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0638051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0638634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0639221Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0640061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0640723Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0641325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0641916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0642510Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0643103Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 6%] 2025-07-17T09:17:08.0643699Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0644295Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0644886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0645480Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0646072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0646653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0647396Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0648024Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0648620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0649211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0649791Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0650375Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0650969Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0651552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0652141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0652730Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0653324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0653916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0654697Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0655357Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0655949Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0656544Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0657131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0657714Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0658308Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0658892Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0659478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0660070Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0660545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0061s] [ 7%] 2025-07-17T09:17:08.0661017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 7%] 2025-07-17T09:17:08.0661602Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 7%] 2025-07-17T09:17:08.0662119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 7%] 2025-07-17T09:17:08.0662635Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0028s] [ 7%] 2025-07-17T09:17:08.0663106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 7%] 2025-07-17T09:17:08.0663582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0664046Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 7%] 2025-07-17T09:17:08.0664516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0047s] [ 7%] 2025-07-17T09:17:08.0664975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 7%] 2025-07-17T09:17:08.0665442Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0665898Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 7%] 2025-07-17T09:17:08.0666361Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0028s] [ 7%] 2025-07-17T09:17:08.0666829Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 7%] 2025-07-17T09:17:08.0667287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0667750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 7%] 2025-07-17T09:17:08.0668218Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0050s] [ 7%] 2025-07-17T09:17:08.0668847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0669359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 7%] 2025-07-17T09:17:08.0669886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 7%] 2025-07-17T09:17:08.0670363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0670834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0671305Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 7%] 2025-07-17T09:17:08.0671774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 7%] 2025-07-17T09:17:08.0672249Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0052s] [ 7%] 2025-07-17T09:17:08.0672709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0673170Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 7%] 2025-07-17T09:17:08.0673629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 7%] 2025-07-17T09:17:08.0674151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 7%] 2025-07-17T09:17:08.0674622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 7%] 2025-07-17T09:17:08.0675083Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 7%] 2025-07-17T09:17:08.0675545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 7%] 2025-07-17T09:17:08.0676144Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0676683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0677207Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 7%] 2025-07-17T09:17:08.0677671Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 7%] 2025-07-17T09:17:08.0678146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0678613Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0679084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 7%] 2025-07-17T09:17:08.0679550Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 7%] 2025-07-17T09:17:08.0680021Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0680486Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0680946Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 7%] 2025-07-17T09:17:08.0681410Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 7%] 2025-07-17T09:17:08.0681872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0682344Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 7%] 2025-07-17T09:17:08.0682805Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 7%] 2025-07-17T09:17:08.0683392Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 7%] 2025-07-17T09:17:08.0684048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0684685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0685277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0685858Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0686453Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0687044Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0687626Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0688212Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0688791Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0689373Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0689958Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0690643Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0691281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0691917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0692506Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0693084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0693669Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0694268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0694853Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0695445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0696046Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0696631Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0697220Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0697911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0698551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0699185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0699765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0700345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0700925Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0701519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0702109Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0702685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0703275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0703857Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0704445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0705133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0705778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0706430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0707014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0707600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 7%] 2025-07-17T09:17:08.0708185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 8%] 2025-07-17T09:17:08.0708765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 8%] 2025-07-17T09:17:08.0709351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 8%] 2025-07-17T09:17:08.0709933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 8%] 2025-07-17T09:17:08.0710516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 8%] 2025-07-17T09:17:08.0711096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 8%] 2025-07-17T09:17:08.0711671Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 8%] 2025-07-17T09:17:08.0712356Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 8%] 2025-07-17T09:17:08.0712894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0075s] [ 8%] 2025-07-17T09:17:08.0713416Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0054s] [ 8%] 2025-07-17T09:17:08.0713954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0117s] [ 8%] 2025-07-17T09:17:08.0714430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0097s] [ 8%] 2025-07-17T09:17:08.0714911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0054s] [ 8%] 2025-07-17T09:17:08.0715382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0054s] [ 8%] 2025-07-17T09:17:08.0715861Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0118s] [ 8%] 2025-07-17T09:17:08.0716337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0097s] [ 8%] 2025-07-17T09:17:08.0716810Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0065s] [ 8%] 2025-07-17T09:17:08.0717285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0053s] [ 8%] 2025-07-17T09:17:08.0717749Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0120s] [ 8%] 2025-07-17T09:17:08.0718217Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0097s] [ 8%] 2025-07-17T09:17:08.0718686Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0054s] [ 8%] 2025-07-17T09:17:08.0719156Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0053s] [ 8%] 2025-07-17T09:17:08.0719803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0117s] [ 8%] 2025-07-17T09:17:08.0720323Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0097s] [ 8%] 2025-07-17T09:17:08.0720802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0081s] [ 8%] 2025-07-17T09:17:08.0721282Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0064s] [ 8%] 2025-07-17T09:17:08.0721759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0141s] [ 8%] 2025-07-17T09:17:08.0722231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0116s] [ 8%] 2025-07-17T09:17:08.0722714Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0063s] [ 8%] 2025-07-17T09:17:08.0723200Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0063s] [ 8%] 2025-07-17T09:17:08.0723673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0139s] [ 8%] 2025-07-17T09:17:08.0724150Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0114s] [ 8%] 2025-07-17T09:17:08.0724631Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0081s] [ 8%] 2025-07-17T09:17:08.0725114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0062s] [ 8%] 2025-07-17T09:17:08.0725577Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0138s] [ 8%] 2025-07-17T09:17:08.0726044Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0114s] [ 8%] 2025-07-17T09:17:08.0726571Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0062s] [ 8%] 2025-07-17T09:17:08.0727139Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0062s] [ 8%] 2025-07-17T09:17:08.0727664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0139s] [ 8%] 2025-07-17T09:17:08.0728131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0113s] [ 8%] 2025-07-17T09:17:08.0728617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0062s] [ 8%] 2025-07-17T09:17:08.0729092Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0062s] [ 8%] 2025-07-17T09:17:08.0729564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0138s] [ 8%] 2025-07-17T09:17:08.0730033Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0113s] [ 8%] 2025-07-17T09:17:08.0730509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0062s] [ 8%] 2025-07-17T09:17:08.0730988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0061s] [ 8%] 2025-07-17T09:17:08.0731461Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0138s] [ 8%] 2025-07-17T09:17:08.0731938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0113s] [ 8%] 2025-07-17T09:17:08.0732410Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0062s] [ 8%] 2025-07-17T09:17:08.0732880Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0061s] [ 8%] 2025-07-17T09:17:08.0733345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0138s] [ 8%] 2025-07-17T09:17:08.0733920Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0113s] [ 8%] 2025-07-17T09:17:08.0734459Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0061s] [ 8%] 2025-07-17T09:17:08.0734980Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0061s] [ 8%] 2025-07-17T09:17:08.0735448Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0139s] [ 8%] 2025-07-17T09:17:08.0735912Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0114s] [ 8%] 2025-07-17T09:17:08.0736388Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0325s] [ 8%] 2025-07-17T09:17:08.0736862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0061s] [ 8%] 2025-07-17T09:17:08.0737326Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0123s] [ 8%] 2025-07-17T09:17:08.0737798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0098s] [ 8%] 2025-07-17T09:17:08.0738270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0055s] [ 8%] 2025-07-17T09:17:08.0738740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0054s] [ 8%] 2025-07-17T09:17:08.0739205Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0122s] [ 8%] 2025-07-17T09:17:08.0739678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0096s] [ 8%] 2025-07-17T09:17:08.0740146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0164s] [ 8%] 2025-07-17T09:17:08.0740609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0057s] [ 8%] 2025-07-17T09:17:08.0741188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0124s] [ 8%] 2025-07-17T09:17:08.0741700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0100s] [ 8%] 2025-07-17T09:17:08.0742221Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0056s] [ 8%] 2025-07-17T09:17:08.0742687Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0055s] [ 8%] 2025-07-17T09:17:08.0743159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0122s] [ 8%] 2025-07-17T09:17:08.0743627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0097s] [ 8%] 2025-07-17T09:17:08.0744098Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0178s] [ 8%] 2025-07-17T09:17:08.0744574Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0063s] [ 8%] 2025-07-17T09:17:08.0745037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0140s] [ 8%] 2025-07-17T09:17:08.0745509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0110s] [ 8%] 2025-07-17T09:17:08.0745981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0063s] [ 8%] 2025-07-17T09:17:08.0746460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0063s] [ 8%] 2025-07-17T09:17:08.0746934Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0141s] [ 8%] 2025-07-17T09:17:08.0747405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0110s] [ 8%] 2025-07-17T09:17:08.0747872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0170s] [ 8%] 2025-07-17T09:17:08.0748490Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0064s] [ 8%] 2025-07-17T09:17:08.0749008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0141s] [ 8%] 2025-07-17T09:17:08.0749468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0114s] [ 8%] 2025-07-17T09:17:08.0749942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0063s] [ 8%] 2025-07-17T09:17:08.0750415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0063s] [ 8%] 2025-07-17T09:17:08.0750879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0139s] [ 8%] 2025-07-17T09:17:08.0751345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0110s] [ 8%] 2025-07-17T09:17:08.0751831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0062s] [ 8%] 2025-07-17T09:17:08.0752304Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0062s] [ 8%] 2025-07-17T09:17:08.0752771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0140s] [ 8%] 2025-07-17T09:17:08.0753239Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0110s] [ 8%] 2025-07-17T09:17:08.0753711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0063s] [ 8%] 2025-07-17T09:17:08.0754245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0063s] [ 8%] 2025-07-17T09:17:08.0754717Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0141s] [ 8%] 2025-07-17T09:17:08.0755392Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0110s] [ 8%] 2025-07-17T09:17:08.0755986Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0063s] [ 8%] 2025-07-17T09:17:08.0756509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0062s] [ 8%] 2025-07-17T09:17:08.0756969Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0141s] [ 8%] 2025-07-17T09:17:08.0757430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0116s] [ 8%] 2025-07-17T09:17:08.0757898Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0069s] [ 8%] 2025-07-17T09:17:08.0758364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0067s] [ 8%] 2025-07-17T09:17:08.0758831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0146s] [ 8%] 2025-07-17T09:17:08.0759296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0116s] [ 8%] 2025-07-17T09:17:08.0759768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0087s] [ 8%] 2025-07-17T09:17:08.0760240Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0183s] [ 8%] 2025-07-17T09:17:08.0760707Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0167s] [ 8%] 2025-07-17T09:17:08.0761175Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0108s] [ 8%] 2025-07-17T09:17:08.0761648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0052s] [ 8%] 2025-07-17T09:17:08.0762120Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0050s] [ 8%] 2025-07-17T09:17:08.0762723Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0122s] [ 8%] 2025-07-17T09:17:08.0763243Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0097s] [ 8%] 2025-07-17T09:17:08.0763762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0080s] [ 8%] 2025-07-17T09:17:08.0764225Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0050s] [ 8%] 2025-07-17T09:17:08.0764695Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0123s] [ 8%] 2025-07-17T09:17:08.0765156Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0100s] [ 8%] 2025-07-17T09:17:08.0765626Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0050s] [ 9%] 2025-07-17T09:17:08.0766098Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0050s] [ 9%] 2025-07-17T09:17:08.0766560Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0122s] [ 9%] 2025-07-17T09:17:08.0767028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0099s] [ 9%] 2025-07-17T09:17:08.0767496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0077s] [ 9%] 2025-07-17T09:17:08.0767971Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0061s] [ 9%] 2025-07-17T09:17:08.0768435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0148s] [ 9%] 2025-07-17T09:17:08.0768906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0116s] [ 9%] 2025-07-17T09:17:08.0769381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0059s] [ 9%] 2025-07-17T09:17:08.0769963Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0058s] [ 9%] 2025-07-17T09:17:08.0770494Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0144s] [ 9%] 2025-07-17T09:17:08.0771016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0115s] [ 9%] 2025-07-17T09:17:08.0771489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0069s] [ 9%] 2025-07-17T09:17:08.0771955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0058s] [ 9%] 2025-07-17T09:17:08.0772423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0144s] [ 9%] 2025-07-17T09:17:08.0772891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0114s] [ 9%] 2025-07-17T09:17:08.0773366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0058s] [ 9%] 2025-07-17T09:17:08.0773835Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0774301Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0143s] [ 9%] 2025-07-17T09:17:08.0774773Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0114s] [ 9%] 2025-07-17T09:17:08.0775243Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0058s] [ 9%] 2025-07-17T09:17:08.0775718Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0776189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0145s] [ 9%] 2025-07-17T09:17:08.0776653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0115s] [ 9%] 2025-07-17T09:17:08.0777292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0058s] [ 9%] 2025-07-17T09:17:08.0777809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0778292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0143s] [ 9%] 2025-07-17T09:17:08.0778765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0114s] [ 9%] 2025-07-17T09:17:08.0779235Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0058s] [ 9%] 2025-07-17T09:17:08.0779709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0780175Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0144s] [ 9%] 2025-07-17T09:17:08.0780645Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0115s] [ 9%] 2025-07-17T09:17:08.0781115Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0058s] [ 9%] 2025-07-17T09:17:08.0781593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0782057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0144s] [ 9%] 2025-07-17T09:17:08.0782526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0114s] [ 9%] 2025-07-17T09:17:08.0782997Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0164s] [ 9%] 2025-07-17T09:17:08.0783463Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0052s] [ 9%] 2025-07-17T09:17:08.0783985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0127s] [ 9%] 2025-07-17T09:17:08.0784561Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0098s] [ 9%] 2025-07-17T09:17:08.0785105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0051s] [ 9%] 2025-07-17T09:17:08.0785575Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0051s] [ 9%] 2025-07-17T09:17:08.0786045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0127s] [ 9%] 2025-07-17T09:17:08.0786511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0098s] [ 9%] 2025-07-17T09:17:08.0786973Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0157s] [ 9%] 2025-07-17T09:17:08.0787435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0052s] [ 9%] 2025-07-17T09:17:08.0787895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0127s] [ 9%] 2025-07-17T09:17:08.0788359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0109s] [ 9%] 2025-07-17T09:17:08.0788824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0052s] [ 9%] 2025-07-17T09:17:08.0789294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0051s] [ 9%] 2025-07-17T09:17:08.0789755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0129s] [ 9%] 2025-07-17T09:17:08.0790217Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0099s] [ 9%] 2025-07-17T09:17:08.0790687Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0178s] [ 9%] 2025-07-17T09:17:08.0791282Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0060s] [ 9%] 2025-07-17T09:17:08.0791804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0147s] [ 9%] 2025-07-17T09:17:08.0792317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0112s] [ 9%] 2025-07-17T09:17:08.0792792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0793271Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0793738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0146s] [ 9%] 2025-07-17T09:17:08.0794272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0112s] [ 9%] 2025-07-17T09:17:08.0794738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0169s] [ 9%] 2025-07-17T09:17:08.0795210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0058s] [ 9%] 2025-07-17T09:17:08.0795669Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0147s] [ 9%] 2025-07-17T09:17:08.0796131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0113s] [ 9%] 2025-07-17T09:17:08.0796606Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0797067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0056s] [ 9%] 2025-07-17T09:17:08.0797530Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0146s] [ 9%] 2025-07-17T09:17:08.0797987Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0113s] [ 9%] 2025-07-17T09:17:08.0798586Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0799119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0799648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0146s] [ 9%] 2025-07-17T09:17:08.0800114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0112s] [ 9%] 2025-07-17T09:17:08.0800596Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0801072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0801538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0147s] [ 9%] 2025-07-17T09:17:08.0802008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0111s] [ 9%] 2025-07-17T09:17:08.0802469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0802936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0803395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0146s] [ 9%] 2025-07-17T09:17:08.0803856Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0113s] [ 9%] 2025-07-17T09:17:08.0804329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0058s] [ 9%] 2025-07-17T09:17:08.0804791Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0057s] [ 9%] 2025-07-17T09:17:08.0805256Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0147s] [ 9%] 2025-07-17T09:17:08.0805905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0112s] [ 9%] 2025-07-17T09:17:08.0806432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 9%] 2025-07-17T09:17:08.0806900Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 9%] 2025-07-17T09:17:08.0807363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 9%] 2025-07-17T09:17:08.0807830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0051s] [ 9%] 2025-07-17T09:17:08.0808299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 9%] 2025-07-17T09:17:08.0808766Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 9%] 2025-07-17T09:17:08.0809234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 9%] 2025-07-17T09:17:08.0809700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0050s] [ 9%] 2025-07-17T09:17:08.0810167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0042s] [ 9%] 2025-07-17T09:17:08.0810626Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 9%] 2025-07-17T09:17:08.0811094Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0059s] [ 9%] 2025-07-17T09:17:08.0811554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0050s] [ 9%] 2025-07-17T09:17:08.0812023Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 9%] 2025-07-17T09:17:08.0812545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 9%] 2025-07-17T09:17:08.0813113Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0059s] [ 9%] 2025-07-17T09:17:08.0813645Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0049s] [ 9%] 2025-07-17T09:17:08.0814114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0071s] [ 9%] 2025-07-17T09:17:08.0814586Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 9%] 2025-07-17T09:17:08.0815051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0082s] [ 9%] 2025-07-17T09:17:08.0815515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0067s] [ 9%] 2025-07-17T09:17:08.0815983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 9%] 2025-07-17T09:17:08.0816456Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 9%] 2025-07-17T09:17:08.0816925Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0082s] [ 9%] 2025-07-17T09:17:08.0817391Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0066s] [ 9%] 2025-07-17T09:17:08.0817855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0069s] [ 9%] 2025-07-17T09:17:08.0818320Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0041s] [ 9%] 2025-07-17T09:17:08.0818783Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0082s] [ 9%] 2025-07-17T09:17:08.0819240Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0068s] [ 9%] 2025-07-17T09:17:08.0819771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 9%] 2025-07-17T09:17:08.0820340Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0040s] [ 9%] 2025-07-17T09:17:08.0820849Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0081s] [ 9%] 2025-07-17T09:17:08.0821314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0067s] [ 9%] 2025-07-17T09:17:08.0821786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 10%] 2025-07-17T09:17:08.0822255Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 10%] 2025-07-17T09:17:08.0822716Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0082s] [ 10%] 2025-07-17T09:17:08.0823184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0067s] [ 10%] 2025-07-17T09:17:08.0823661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 10%] 2025-07-17T09:17:08.0824126Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 10%] 2025-07-17T09:17:08.0824593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0082s] [ 10%] 2025-07-17T09:17:08.0825057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0067s] [ 10%] 2025-07-17T09:17:08.0825524Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 10%] 2025-07-17T09:17:08.0825984Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0040s] [ 10%] 2025-07-17T09:17:08.0826445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0081s] [ 10%] 2025-07-17T09:17:08.0827031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0067s] [ 10%] 2025-07-17T09:17:08.0827548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 10%] 2025-07-17T09:17:08.0828067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0040s] [ 10%] 2025-07-17T09:17:08.0828525Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0081s] [ 10%] 2025-07-17T09:17:08.0828995Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0066s] [ 10%] 2025-07-17T09:17:08.0829460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0147s] [ 10%] 2025-07-17T09:17:08.0829927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 10%] 2025-07-17T09:17:08.0830395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0064s] [ 10%] 2025-07-17T09:17:08.0830859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0052s] [ 10%] 2025-07-17T09:17:08.0831332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 10%] 2025-07-17T09:17:08.0831793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 10%] 2025-07-17T09:17:08.0832257Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0064s] [ 10%] 2025-07-17T09:17:08.0832719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0053s] [ 10%] 2025-07-17T09:17:08.0833181Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0133s] [ 10%] 2025-07-17T09:17:08.0833642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 10%] 2025-07-17T09:17:08.0834282Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0064s] [ 10%] 2025-07-17T09:17:08.0834812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0054s] [ 10%] 2025-07-17T09:17:08.0835326Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 10%] 2025-07-17T09:17:08.0835793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 10%] 2025-07-17T09:17:08.0836252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0064s] [ 10%] 2025-07-17T09:17:08.0836711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0052s] [ 10%] 2025-07-17T09:17:08.0837179Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0160s] [ 10%] 2025-07-17T09:17:08.0837641Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0043s] [ 10%] 2025-07-17T09:17:08.0838108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0085s] [ 10%] 2025-07-17T09:17:08.0838572Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0069s] [ 10%] 2025-07-17T09:17:08.0839045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0839515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0839983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0085s] [ 10%] 2025-07-17T09:17:08.0840447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0070s] [ 10%] 2025-07-17T09:17:08.0840902Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0153s] [ 10%] 2025-07-17T09:17:08.0841491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0043s] [ 10%] 2025-07-17T09:17:08.0842018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0085s] [ 10%] 2025-07-17T09:17:08.0842532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0069s] [ 10%] 2025-07-17T09:17:08.0842991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0843460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0843923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0084s] [ 10%] 2025-07-17T09:17:08.0844387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0069s] [ 10%] 2025-07-17T09:17:08.0844861Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0845331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0845803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0085s] [ 10%] 2025-07-17T09:17:08.0846260Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0070s] [ 10%] 2025-07-17T09:17:08.0846732Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0847201Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0847667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0086s] [ 10%] 2025-07-17T09:17:08.0848132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0070s] [ 10%] 2025-07-17T09:17:08.0848698Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0849213Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0043s] [ 10%] 2025-07-17T09:17:08.0849722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0085s] [ 10%] 2025-07-17T09:17:08.0850190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0069s] [ 10%] 2025-07-17T09:17:08.0850653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0851117Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0851574Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0085s] [ 10%] 2025-07-17T09:17:08.0852033Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0069s] [ 10%] 2025-07-17T09:17:08.0852497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 10%] 2025-07-17T09:17:08.0852957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 10%] 2025-07-17T09:17:08.0853420Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 10%] 2025-07-17T09:17:08.0853888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0043s] [ 10%] 2025-07-17T09:17:08.0854355Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 10%] 2025-07-17T09:17:08.0854823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 10%] 2025-07-17T09:17:08.0855290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 10%] 2025-07-17T09:17:08.0855870Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0856379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0053s] [ 10%] 2025-07-17T09:17:08.0856887Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 10%] 2025-07-17T09:17:08.0857351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 10%] 2025-07-17T09:17:08.0857808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0043s] [ 10%] 2025-07-17T09:17:08.0858270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 10%] 2025-07-17T09:17:08.0858727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 10%] 2025-07-17T09:17:08.0859190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 10%] 2025-07-17T09:17:08.0859648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0860117Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 10%] 2025-07-17T09:17:08.0860582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 10%] 2025-07-17T09:17:08.0861042Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 10%] 2025-07-17T09:17:08.0861513Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0044s] [ 10%] 2025-07-17T09:17:08.0861978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 10%] 2025-07-17T09:17:08.0862447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 10%] 2025-07-17T09:17:08.0863058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 10%] 2025-07-17T09:17:08.0863698Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0044s] [ 10%] 2025-07-17T09:17:08.0864161Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0038s] [ 10%] 2025-07-17T09:17:08.0864620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 10%] 2025-07-17T09:17:08.0865082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0041s] [ 10%] 2025-07-17T09:17:08.0865549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0045s] [ 10%] 2025-07-17T09:17:08.0866012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0038s] [ 10%] 2025-07-17T09:17:08.0866470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 10%] 2025-07-17T09:17:08.0866933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0867393Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0045s] [ 10%] 2025-07-17T09:17:08.0867855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 10%] 2025-07-17T09:17:08.0868322Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 10%] 2025-07-17T09:17:08.0868780Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 10%] 2025-07-17T09:17:08.0869246Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0044s] [ 10%] 2025-07-17T09:17:08.0869765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 10%] 2025-07-17T09:17:08.0870348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 10%] 2025-07-17T09:17:08.0870868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 10%] 2025-07-17T09:17:08.0871331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0044s] [ 10%] 2025-07-17T09:17:08.0871799Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 10%] 2025-07-17T09:17:08.0872265Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 10%] 2025-07-17T09:17:08.0872730Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0041s] [ 10%] 2025-07-17T09:17:08.0873186Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0044s] [ 10%] 2025-07-17T09:17:08.0873654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 10%] 2025-07-17T09:17:08.0874187Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 10%] 2025-07-17T09:17:08.0874647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 10%] 2025-07-17T09:17:08.0875107Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0044s] [ 10%] 2025-07-17T09:17:08.0875705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 10%] 2025-07-17T09:17:08.0876294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 10%] 2025-07-17T09:17:08.0876891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 10%] 2025-07-17T09:17:08.0877665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0015s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 10%] 2025-07-17T09:17:08.0878317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0878906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0879496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0880082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0880663Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0881250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0881826Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0882408Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0882989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0883567Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0884145Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0885015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0885705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0886305Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0886897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0887484Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0888072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0888668Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0889257Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0889842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0890441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0891021Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0891599Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0892341Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0892978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0893562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0894143Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0894734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0895326Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0895918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0896509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0897091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0897688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0898281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0898865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0899620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0900270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0900856Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0901441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0902022Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0902602Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0903184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0903764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0904351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0904828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0069s] [ 11%] 2025-07-17T09:17:08.0905298Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 11%] 2025-07-17T09:17:08.0905761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 11%] 2025-07-17T09:17:08.0906279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 11%] 2025-07-17T09:17:08.0906868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 11%] 2025-07-17T09:17:08.0907393Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 11%] 2025-07-17T09:17:08.0907866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 11%] 2025-07-17T09:17:08.0908333Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 11%] 2025-07-17T09:17:08.0908801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0050s] [ 11%] 2025-07-17T09:17:08.0909265Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 11%] 2025-07-17T09:17:08.0909727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 11%] 2025-07-17T09:17:08.0910192Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 11%] 2025-07-17T09:17:08.0910656Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 11%] 2025-07-17T09:17:08.0911121Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 11%] 2025-07-17T09:17:08.0911581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 11%] 2025-07-17T09:17:08.0912048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 11%] 2025-07-17T09:17:08.0912520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 11%] 2025-07-17T09:17:08.0912988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 11%] 2025-07-17T09:17:08.0913508Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 11%] 2025-07-17T09:17:08.0914145Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 11%] 2025-07-17T09:17:08.0914688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 11%] 2025-07-17T09:17:08.0915158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 11%] 2025-07-17T09:17:08.0915646Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 11%] 2025-07-17T09:17:08.0916115Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 11%] 2025-07-17T09:17:08.0916584Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 11%] 2025-07-17T09:17:08.0917048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 11%] 2025-07-17T09:17:08.0917519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 11%] 2025-07-17T09:17:08.0917983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0041s] [ 11%] 2025-07-17T09:17:08.0918447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 11%] 2025-07-17T09:17:08.0918911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 11%] 2025-07-17T09:17:08.0919380Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 11%] 2025-07-17T09:17:08.0919846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0041s] [ 11%] 2025-07-17T09:17:08.0920318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 11%] 2025-07-17T09:17:08.0920844Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 11%] 2025-07-17T09:17:08.0921423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 11%] 2025-07-17T09:17:08.0921957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 11%] 2025-07-17T09:17:08.0922435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 11%] 2025-07-17T09:17:08.0922905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 11%] 2025-07-17T09:17:08.0923382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 11%] 2025-07-17T09:17:08.0923852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 11%] 2025-07-17T09:17:08.0924315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 11%] 2025-07-17T09:17:08.0924783Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 11%] 2025-07-17T09:17:08.0925241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0041s] [ 11%] 2025-07-17T09:17:08.0925703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0041s] [ 11%] 2025-07-17T09:17:08.0926165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 11%] 2025-07-17T09:17:08.0926636Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 11%] 2025-07-17T09:17:08.0927105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 11%] 2025-07-17T09:17:08.0927563Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0040s] [ 11%] 2025-07-17T09:17:08.0928268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0928915Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0929555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0930140Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0930737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0931330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0931917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0932504Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0933085Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0933669Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0934251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0934834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0935528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0936172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0936800Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0937386Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0937975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0938570Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0939160Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0939745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0940337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0940923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0941515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0942102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 11%] 2025-07-17T09:17:08.0942745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0943431Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0944059Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0944645Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0945232Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0945812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0946398Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0946980Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0947575Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0948165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0948750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0949339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0949982Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0950698Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0951339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0951928Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0952519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0953100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0953683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0954360Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0954947Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0955536Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0956118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0956699Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0957236Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0056s] [ 12%] 2025-07-17T09:17:08.0957839Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 12%] 2025-07-17T09:17:08.0958375Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 12%] 2025-07-17T09:17:08.0958835Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 12%] 2025-07-17T09:17:08.0959311Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 12%] 2025-07-17T09:17:08.0959775Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 12%] 2025-07-17T09:17:08.0960246Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0960708Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0961167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0039s] [ 12%] 2025-07-17T09:17:08.0961629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 12%] 2025-07-17T09:17:08.0962087Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0962547Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0963008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 12%] 2025-07-17T09:17:08.0963472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 12%] 2025-07-17T09:17:08.0963932Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0964440Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0965016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0059s] [ 12%] 2025-07-17T09:17:08.0965532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 12%] 2025-07-17T09:17:08.0966004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 12%] 2025-07-17T09:17:08.0966470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 12%] 2025-07-17T09:17:08.0966944Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 12%] 2025-07-17T09:17:08.0967419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 12%] 2025-07-17T09:17:08.0967880Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 12%] 2025-07-17T09:17:08.0968349Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 12%] 2025-07-17T09:17:08.0968805Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0052s] [ 12%] 2025-07-17T09:17:08.0969269Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 12%] 2025-07-17T09:17:08.0969721Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 12%] 2025-07-17T09:17:08.0970183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 12%] 2025-07-17T09:17:08.0970650Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 12%] 2025-07-17T09:17:08.0971110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0971638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 12%] 2025-07-17T09:17:08.0972202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 12%] 2025-07-17T09:17:08.0972726Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0973193Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0973670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 12%] 2025-07-17T09:17:08.0974142Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 12%] 2025-07-17T09:17:08.0974612Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0975080Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0975543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 12%] 2025-07-17T09:17:08.0976010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 12%] 2025-07-17T09:17:08.0976469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 12%] 2025-07-17T09:17:08.0976931Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 12%] 2025-07-17T09:17:08.0977396Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 12%] 2025-07-17T09:17:08.0977851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 12%] 2025-07-17T09:17:08.0978319Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0978828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 12%] 2025-07-17T09:17:08.0980765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 12%] 2025-07-17T09:17:08.0981274Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 12%] 2025-07-17T09:17:08.0981875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0982471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0983158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0983761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0984353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0984942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0985528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0986112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0986701Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0987283Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0988075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0988716Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0989348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0989937Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0990515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0991099Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0991693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0992278Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0992863Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0993446Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0994142Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0994735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0995447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0996103Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0996743Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0997336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0997916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0998490Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0999075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.0999649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1000228Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1000806Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1001385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1001978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1002667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1003306Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1003948Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1004535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1005128Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1005708Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1006291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1006873Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1007450Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1008027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 12%] 2025-07-17T09:17:08.1008606Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1009190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1009848Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1010526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1011063Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0074s] [ 13%] 2025-07-17T09:17:08.1011533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 13%] 2025-07-17T09:17:08.1012016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0062s] [ 13%] 2025-07-17T09:17:08.1012491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0062s] [ 13%] 2025-07-17T09:17:08.1012966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 13%] 2025-07-17T09:17:08.1013444Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 13%] 2025-07-17T09:17:08.1013916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0062s] [ 13%] 2025-07-17T09:17:08.1014393Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0063s] [ 13%] 2025-07-17T09:17:08.1014859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0063s] [ 13%] 2025-07-17T09:17:08.1015330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 13%] 2025-07-17T09:17:08.1015801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0063s] [ 13%] 2025-07-17T09:17:08.1016263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0062s] [ 13%] 2025-07-17T09:17:08.1016733Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 13%] 2025-07-17T09:17:08.1017321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 13%] 2025-07-17T09:17:08.1017850Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0060s] [ 13%] 2025-07-17T09:17:08.1018364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0061s] [ 13%] 2025-07-17T09:17:08.1018851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 13%] 2025-07-17T09:17:08.1019329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 13%] 2025-07-17T09:17:08.1019799Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0066s] [ 13%] 2025-07-17T09:17:08.1020270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0062s] [ 13%] 2025-07-17T09:17:08.1020744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 13%] 2025-07-17T09:17:08.1021227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 13%] 2025-07-17T09:17:08.1021700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0066s] [ 13%] 2025-07-17T09:17:08.1022176Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0063s] [ 13%] 2025-07-17T09:17:08.1022649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0039s] [ 13%] 2025-07-17T09:17:08.1023113Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 13%] 2025-07-17T09:17:08.1023585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0065s] [ 13%] 2025-07-17T09:17:08.1024102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0063s] [ 13%] 2025-07-17T09:17:08.1024678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0038s] [ 13%] 2025-07-17T09:17:08.1025203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 13%] 2025-07-17T09:17:08.1025675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0066s] [ 13%] 2025-07-17T09:17:08.1026148Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0062s] [ 13%] 2025-07-17T09:17:08.1026621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 13%] 2025-07-17T09:17:08.1027100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 13%] 2025-07-17T09:17:08.1027568Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0065s] [ 13%] 2025-07-17T09:17:08.1028041Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0062s] [ 13%] 2025-07-17T09:17:08.1028517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 13%] 2025-07-17T09:17:08.1028994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 13%] 2025-07-17T09:17:08.1029469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0065s] [ 13%] 2025-07-17T09:17:08.1029942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0062s] [ 13%] 2025-07-17T09:17:08.1030418Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0038s] [ 13%] 2025-07-17T09:17:08.1030883Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 13%] 2025-07-17T09:17:08.1031410Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0065s] [ 13%] 2025-07-17T09:17:08.1031971Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0063s] [ 13%] 2025-07-17T09:17:08.1032522Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0038s] [ 13%] 2025-07-17T09:17:08.1032994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 13%] 2025-07-17T09:17:08.1033467Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0065s] [ 13%] 2025-07-17T09:17:08.1034001Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0062s] [ 13%] 2025-07-17T09:17:08.1034604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1035204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1035804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1036396Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1036995Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1037588Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1038179Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1038843Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1039558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1040202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1040814Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1041404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1041994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1042581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1043176Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1043754Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1044352Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1044947Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1045536Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1046178Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1046878Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1047527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1048126Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1048722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1049312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1049894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1050483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1051067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1051648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1052242Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1052820Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1053472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1054177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1054820Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1055412Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1056007Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1056609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1057206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1057798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1058391Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1058977Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1059567Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1060157Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1060794Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1061507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1062151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1062750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1063346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 13%] 2025-07-17T09:17:08.1063826Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0054s] [ 13%] 2025-07-17T09:17:08.1064301Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 13%] 2025-07-17T09:17:08.1064772Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 13%] 2025-07-17T09:17:08.1065244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0055s] [ 13%] 2025-07-17T09:17:08.1065722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 13%] 2025-07-17T09:17:08.1066194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 13%] 2025-07-17T09:17:08.1066667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 13%] 2025-07-17T09:17:08.1067136Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0055s] [ 13%] 2025-07-17T09:17:08.1067606Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0043s] [ 13%] 2025-07-17T09:17:08.1068182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 13%] 2025-07-17T09:17:08.1068705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0060s] [ 13%] 2025-07-17T09:17:08.1069220Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0055s] [ 13%] 2025-07-17T09:17:08.1069688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 13%] 2025-07-17T09:17:08.1070160Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 13%] 2025-07-17T09:17:08.1070626Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0060s] [ 13%] 2025-07-17T09:17:08.1071096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0055s] [ 13%] 2025-07-17T09:17:08.1071573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 14%] 2025-07-17T09:17:08.1072051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 14%] 2025-07-17T09:17:08.1072526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0066s] [ 14%] 2025-07-17T09:17:08.1072990Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0060s] [ 14%] 2025-07-17T09:17:08.1073467Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1073990Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1074474Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0067s] [ 14%] 2025-07-17T09:17:08.1074942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0060s] [ 14%] 2025-07-17T09:17:08.1075540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0045s] [ 14%] 2025-07-17T09:17:08.1076088Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1076613Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0066s] [ 14%] 2025-07-17T09:17:08.1077084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0060s] [ 14%] 2025-07-17T09:17:08.1077554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1078026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1078488Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0066s] [ 14%] 2025-07-17T09:17:08.1078969Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0060s] [ 14%] 2025-07-17T09:17:08.1079451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1079920Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 14%] 2025-07-17T09:17:08.1080389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0067s] [ 14%] 2025-07-17T09:17:08.1080855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0060s] [ 14%] 2025-07-17T09:17:08.1081338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1081809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1082341Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0066s] [ 14%] 2025-07-17T09:17:08.1082923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0060s] [ 14%] 2025-07-17T09:17:08.1083443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 14%] 2025-07-17T09:17:08.1083916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1084386Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0066s] [ 14%] 2025-07-17T09:17:08.1084851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0060s] [ 14%] 2025-07-17T09:17:08.1085320Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1085789Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1086262Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0067s] [ 14%] 2025-07-17T09:17:08.1086728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0059s] [ 14%] 2025-07-17T09:17:08.1087330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1087923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1088523Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1089115Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1089759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1090489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1091139Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1091736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1092331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1092914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1093507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1094092Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1094689Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1095284Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1095870Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1096457Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1097103Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1098171Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1098820Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1099417Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1100019Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1100613Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1101210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1101803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1102393Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1102987Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1103569Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1104159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1104862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1105498Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1106152Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1106741Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1107338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1107935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1108528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1109124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1109719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1110332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1110925Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1111511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1112209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1112850Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1113485Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1114126Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1114716Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1115310Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1115906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1116492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 14%] 2025-07-17T09:17:08.1116967Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0051s] [ 14%] 2025-07-17T09:17:08.1117439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 14%] 2025-07-17T09:17:08.1117910Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 14%] 2025-07-17T09:17:08.1118372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 14%] 2025-07-17T09:17:08.1118845Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0028s] [ 14%] 2025-07-17T09:17:08.1119465Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 14%] 2025-07-17T09:17:08.1119999Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 14%] 2025-07-17T09:17:08.1120550Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1121018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1121495Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 14%] 2025-07-17T09:17:08.1121957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 14%] 2025-07-17T09:17:08.1122430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 14%] 2025-07-17T09:17:08.1122902Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0027s] [ 14%] 2025-07-17T09:17:08.1123367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 14%] 2025-07-17T09:17:08.1123832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 14%] 2025-07-17T09:17:08.1124293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 14%] 2025-07-17T09:17:08.1124773Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 14%] 2025-07-17T09:17:08.1125244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 14%] 2025-07-17T09:17:08.1125716Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 14%] 2025-07-17T09:17:08.1126184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0046s] [ 14%] 2025-07-17T09:17:08.1126769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 14%] 2025-07-17T09:17:08.1127298Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 14%] 2025-07-17T09:17:08.1127813Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 14%] 2025-07-17T09:17:08.1128285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0047s] [ 14%] 2025-07-17T09:17:08.1128752Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0039s] [ 14%] 2025-07-17T09:17:08.1129224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 14%] 2025-07-17T09:17:08.1129703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 14%] 2025-07-17T09:17:08.1130167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0046s] [ 14%] 2025-07-17T09:17:08.1130649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 14%] 2025-07-17T09:17:08.1131119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 14%] 2025-07-17T09:17:08.1131589Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 14%] 2025-07-17T09:17:08.1132051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0046s] [ 14%] 2025-07-17T09:17:08.1132536Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 14%] 2025-07-17T09:17:08.1133013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 14%] 2025-07-17T09:17:08.1133480Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 14%] 2025-07-17T09:17:08.1134071Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0046s] [ 14%] 2025-07-17T09:17:08.1134594Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1135114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1135580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 15%] 2025-07-17T09:17:08.1136057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0047s] [ 15%] 2025-07-17T09:17:08.1136526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 15%] 2025-07-17T09:17:08.1136990Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1137456Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 15%] 2025-07-17T09:17:08.1137917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0046s] [ 15%] 2025-07-17T09:17:08.1138387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1138847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1139316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 15%] 2025-07-17T09:17:08.1139784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0046s] [ 15%] 2025-07-17T09:17:08.1140385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1141056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1141758Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1142430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1143029Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1143618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1144214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1144799Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1145389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1145977Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1146554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1147143Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1147722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1148378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1149061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1149691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1150287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1150876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1151467Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1152059Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1152650Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1153253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1153841Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1154502Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1155096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1155744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1156462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1157098Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1157692Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1158279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1158860Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1159444Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1160041Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1160634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1161222Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1161807Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1162401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1163039Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1163748Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1164392Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1164979Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1165568Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1166144Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1166727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1167316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1167902Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1168488Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1169069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1169546Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0072s] [ 15%] 2025-07-17T09:17:08.1170014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 15%] 2025-07-17T09:17:08.1170583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 15%] 2025-07-17T09:17:08.1171120Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 15%] 2025-07-17T09:17:08.1171643Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 15%] 2025-07-17T09:17:08.1172118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 15%] 2025-07-17T09:17:08.1172582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1173051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 15%] 2025-07-17T09:17:08.1173532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0052s] [ 15%] 2025-07-17T09:17:08.1173996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 15%] 2025-07-17T09:17:08.1174461Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 15%] 2025-07-17T09:17:08.1174919Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 15%] 2025-07-17T09:17:08.1175389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 15%] 2025-07-17T09:17:08.1175854Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 15%] 2025-07-17T09:17:08.1176322Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1176787Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 15%] 2025-07-17T09:17:08.1177251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0054s] [ 15%] 2025-07-17T09:17:08.1177840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1178353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 15%] 2025-07-17T09:17:08.1178869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 15%] 2025-07-17T09:17:08.1179351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1179829Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1180304Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 15%] 2025-07-17T09:17:08.1180769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 15%] 2025-07-17T09:17:08.1181242Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0054s] [ 15%] 2025-07-17T09:17:08.1181703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1182170Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 15%] 2025-07-17T09:17:08.1182627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 15%] 2025-07-17T09:17:08.1183104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1183571Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1184030Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 15%] 2025-07-17T09:17:08.1184493Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 15%] 2025-07-17T09:17:08.1185077Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1185599Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1186108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 15%] 2025-07-17T09:17:08.1186576Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 15%] 2025-07-17T09:17:08.1187059Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1187529Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1187996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 15%] 2025-07-17T09:17:08.1188460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 15%] 2025-07-17T09:17:08.1188928Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1189390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1189852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 15%] 2025-07-17T09:17:08.1190316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 15%] 2025-07-17T09:17:08.1190780Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1191246Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 15%] 2025-07-17T09:17:08.1191703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 15%] 2025-07-17T09:17:08.1192285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 15%] 2025-07-17T09:17:08.1192930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1193580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1194281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1194870Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1195466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1196049Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1196641Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1197228Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 15%] 2025-07-17T09:17:08.1197810Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1198400Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1198974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1199723Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1200375Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1201016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1201599Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1202175Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1202772Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1203364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1203950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1204535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1205132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1205719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1206312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1207002Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1207638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1208276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1208857Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1209436Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1210016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1210602Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1211183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1211759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1212347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1212932Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1213521Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1214297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1214988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1215638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1216226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1216820Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1217410Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1217999Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1218581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1219160Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1219746Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1220331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1220908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1221543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1222119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0067s] [ 16%] 2025-07-17T09:17:08.1222638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 16%] 2025-07-17T09:17:08.1223102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 16%] 2025-07-17T09:17:08.1223572Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 16%] 2025-07-17T09:17:08.1224045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 16%] 2025-07-17T09:17:08.1224511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 16%] 2025-07-17T09:17:08.1224981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 16%] 2025-07-17T09:17:08.1225451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 16%] 2025-07-17T09:17:08.1225926Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0052s] [ 16%] 2025-07-17T09:17:08.1226387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 16%] 2025-07-17T09:17:08.1226855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 16%] 2025-07-17T09:17:08.1227317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 16%] 2025-07-17T09:17:08.1227778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 16%] 2025-07-17T09:17:08.1228246Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 16%] 2025-07-17T09:17:08.1228853Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 16%] 2025-07-17T09:17:08.1229372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 16%] 2025-07-17T09:17:08.1229894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0064s] [ 16%] 2025-07-17T09:17:08.1230365Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 16%] 2025-07-17T09:17:08.1230839Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 16%] 2025-07-17T09:17:08.1231302Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 16%] 2025-07-17T09:17:08.1231777Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 16%] 2025-07-17T09:17:08.1232244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 16%] 2025-07-17T09:17:08.1232717Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 16%] 2025-07-17T09:17:08.1233185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 16%] 2025-07-17T09:17:08.1233650Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0061s] [ 16%] 2025-07-17T09:17:08.1234198Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 16%] 2025-07-17T09:17:08.1234681Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 16%] 2025-07-17T09:17:08.1235149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 16%] 2025-07-17T09:17:08.1235615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 16%] 2025-07-17T09:17:08.1236220Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 16%] 2025-07-17T09:17:08.1236742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 16%] 2025-07-17T09:17:08.1237268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 16%] 2025-07-17T09:17:08.1237741Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 16%] 2025-07-17T09:17:08.1238208Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 16%] 2025-07-17T09:17:08.1238679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 16%] 2025-07-17T09:17:08.1239141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 16%] 2025-07-17T09:17:08.1239619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 16%] 2025-07-17T09:17:08.1240088Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 16%] 2025-07-17T09:17:08.1240558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 16%] 2025-07-17T09:17:08.1241026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 16%] 2025-07-17T09:17:08.1241484Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 16%] 2025-07-17T09:17:08.1241946Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 16%] 2025-07-17T09:17:08.1242404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 16%] 2025-07-17T09:17:08.1242865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 16%] 2025-07-17T09:17:08.1243441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 16%] 2025-07-17T09:17:08.1243959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 16%] 2025-07-17T09:17:08.1244474Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 16%] 2025-07-17T09:17:08.1244935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 16%] 2025-07-17T09:17:08.1245543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1246135Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1246722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1247313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1247901Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1248489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1249081Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1249675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1250262Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1250992Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1251664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1252240Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1252829Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1253414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1253991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1254576Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1255165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1255756Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1256348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1256928Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1257523Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1258217Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1258876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1259512Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1260105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1260696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1261275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 16%] 2025-07-17T09:17:08.1261859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1262449Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1263028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1263617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1264196Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1264788Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1265489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1266136Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1266771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1267362Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1267956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1268691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1269325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1270019Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1279685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1280453Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1281154Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1281782Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1282629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1283316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1283973Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1284468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0065s] [ 17%] 2025-07-17T09:17:08.1284944Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0027s] [ 17%] 2025-07-17T09:17:08.1285415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 17%] 2025-07-17T09:17:08.1285879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1286349Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0027s] [ 17%] 2025-07-17T09:17:08.1286828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 17%] 2025-07-17T09:17:08.1287309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1287768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 17%] 2025-07-17T09:17:08.1288238Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0049s] [ 17%] 2025-07-17T09:17:08.1288710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 17%] 2025-07-17T09:17:08.1289183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1289695Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0028s] [ 17%] 2025-07-17T09:17:08.1290266Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0027s] [ 17%] 2025-07-17T09:17:08.1290773Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 17%] 2025-07-17T09:17:08.1291239Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0028s] [ 17%] 2025-07-17T09:17:08.1291705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0028s] [ 17%] 2025-07-17T09:17:08.1292182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0052s] [ 17%] 2025-07-17T09:17:08.1292653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1293114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 17%] 2025-07-17T09:17:08.1293580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 17%] 2025-07-17T09:17:08.1294048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1294521Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1294982Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 17%] 2025-07-17T09:17:08.1295452Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 17%] 2025-07-17T09:17:08.1295922Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0052s] [ 17%] 2025-07-17T09:17:08.1296381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1296898Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 17%] 2025-07-17T09:17:08.1297453Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 17%] 2025-07-17T09:17:08.1297997Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1298459Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1298933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 17%] 2025-07-17T09:17:08.1299395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 17%] 2025-07-17T09:17:08.1299859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1300324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1300784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 17%] 2025-07-17T09:17:08.1301248Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 17%] 2025-07-17T09:17:08.1301720Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1302202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1302675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 17%] 2025-07-17T09:17:08.1303140Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 17%] 2025-07-17T09:17:08.1303604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 17%] 2025-07-17T09:17:08.1304119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1304693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 17%] 2025-07-17T09:17:08.1305236Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 17%] 2025-07-17T09:17:08.1305851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1306414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 17%] 2025-07-17T09:17:08.1311338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 17%] 2025-07-17T09:17:08.1311908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 17%] 2025-07-17T09:17:08.1312513Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1313108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1313702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1314381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1315028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1315680Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1316293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1317090Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1317821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1318548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1319225Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1319834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1320477Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1321087Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1321719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1322349Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1322976Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1323624Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1324234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1325114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1325821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1326483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1327163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1327796Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1328438Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1329086Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1329679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1330312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1330917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1331548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1332175Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1353415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1354274Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1354945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1355589Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1356233Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1356843Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1357477Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1358080Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1358713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1359356Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1359953Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1360585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1361242Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1361987Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1362670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1363271Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 17%] 2025-07-17T09:17:08.1363908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1364413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0056s] [ 18%] 2025-07-17T09:17:08.1364938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 18%] 2025-07-17T09:17:08.1365474Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 18%] 2025-07-17T09:17:08.1365962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 18%] 2025-07-17T09:17:08.1366500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 18%] 2025-07-17T09:17:08.1366993Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 18%] 2025-07-17T09:17:08.1367531Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 18%] 2025-07-17T09:17:08.1368013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 18%] 2025-07-17T09:17:08.1368497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 18%] 2025-07-17T09:17:08.1369086Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 18%] 2025-07-17T09:17:08.1369597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 18%] 2025-07-17T09:17:08.1370114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 18%] 2025-07-17T09:17:08.1370585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 18%] 2025-07-17T09:17:08.1371051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 18%] 2025-07-17T09:17:08.1371516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 18%] 2025-07-17T09:17:08.1371978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 18%] 2025-07-17T09:17:08.1372446Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0054s] [ 18%] 2025-07-17T09:17:08.1372911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 18%] 2025-07-17T09:17:08.1373380Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 18%] 2025-07-17T09:17:08.1373844Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 18%] 2025-07-17T09:17:08.1374318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 18%] 2025-07-17T09:17:08.1374791Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 18%] 2025-07-17T09:17:08.1375258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 18%] 2025-07-17T09:17:08.1375722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 18%] 2025-07-17T09:17:08.1376301Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0051s] [ 18%] 2025-07-17T09:17:08.1376825Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 18%] 2025-07-17T09:17:08.1377341Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 18%] 2025-07-17T09:17:08.1377799Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 18%] 2025-07-17T09:17:08.1378273Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 18%] 2025-07-17T09:17:08.1378735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 18%] 2025-07-17T09:17:08.1379201Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 18%] 2025-07-17T09:17:08.1379659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 18%] 2025-07-17T09:17:08.1380133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 18%] 2025-07-17T09:17:08.1380598Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 18%] 2025-07-17T09:17:08.1381057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 18%] 2025-07-17T09:17:08.1381524Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 18%] 2025-07-17T09:17:08.1381995Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 18%] 2025-07-17T09:17:08.1382471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 18%] 2025-07-17T09:17:08.1382932Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 18%] 2025-07-17T09:17:08.1383524Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 18%] 2025-07-17T09:17:08.1384041Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 18%] 2025-07-17T09:17:08.1384550Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 18%] 2025-07-17T09:17:08.1385013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 18%] 2025-07-17T09:17:08.1385474Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 18%] 2025-07-17T09:17:08.1385947Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 18%] 2025-07-17T09:17:08.1386406Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 18%] 2025-07-17T09:17:08.1386878Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0041s] [ 18%] 2025-07-17T09:17:08.1387345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 18%] 2025-07-17T09:17:08.1387938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1388526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1389112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1389700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1390293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1391045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1391702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1392291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1392883Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1393466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1394111Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1394698Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1395280Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1395863Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1396443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1397018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1397609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1398387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1399040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1399627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1400218Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1400808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1401390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1401983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1402566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1403139Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1403721Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1404293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1404876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1405624Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1406256Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1406837Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1407427Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1408014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1408597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1409184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1409785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1410367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1410957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1411541Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1412119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1412896Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1413523Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1414104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1414690Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1415267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1415845Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1416419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 18%] 2025-07-17T09:17:08.1416892Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0053s] [ 18%] 2025-07-17T09:17:08.1417366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 18%] 2025-07-17T09:17:08.1417828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 18%] 2025-07-17T09:17:08.1418298Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 18%] 2025-07-17T09:17:08.1418765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 18%] 2025-07-17T09:17:08.1419240Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 18%] 2025-07-17T09:17:08.1419820Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 18%] 2025-07-17T09:17:08.1420335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 18%] 2025-07-17T09:17:08.1420853Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 18%] 2025-07-17T09:17:08.1421315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 18%] 2025-07-17T09:17:08.1421781Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 18%] 2025-07-17T09:17:08.1422241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 18%] 2025-07-17T09:17:08.1422710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 18%] 2025-07-17T09:17:08.1423166Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 18%] 2025-07-17T09:17:08.1423630Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 18%] 2025-07-17T09:17:08.1424093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 18%] 2025-07-17T09:17:08.1424558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 18%] 2025-07-17T09:17:08.1425027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 18%] 2025-07-17T09:17:08.1425490Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 18%] 2025-07-17T09:17:08.1425959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 19%] 2025-07-17T09:17:08.1426427Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1427016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1427543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 19%] 2025-07-17T09:17:08.1428063Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 19%] 2025-07-17T09:17:08.1428532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 19%] 2025-07-17T09:17:08.1428997Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1429463Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 19%] 2025-07-17T09:17:08.1429924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 19%] 2025-07-17T09:17:08.1430395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1430864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1431322Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 19%] 2025-07-17T09:17:08.1431784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 19%] 2025-07-17T09:17:08.1432259Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1432728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1433193Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 19%] 2025-07-17T09:17:08.1433660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 19%] 2025-07-17T09:17:08.1434335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1434861Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1435411Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 19%] 2025-07-17T09:17:08.1435875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 19%] 2025-07-17T09:17:08.1436349Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 19%] 2025-07-17T09:17:08.1436811Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1437280Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 19%] 2025-07-17T09:17:08.1437744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 19%] 2025-07-17T09:17:08.1438212Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1438676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1439143Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 19%] 2025-07-17T09:17:08.1439608Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 19%] 2025-07-17T09:17:08.1440209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1440797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1441520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1442173Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1442817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1443407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1443994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1444579Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1445158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1445740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1446323Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1446896Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1447479Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1448061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1448700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1449376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1450018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1450619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1451202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1451788Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1452381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1452971Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1453565Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1454146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1454738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1455320Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1455961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1456640Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1457273Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1457860Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1458446Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1459019Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1459614Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1460198Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1460783Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1461369Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1461956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1462542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1463294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1464001Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1464647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1465225Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1465810Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1466384Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1466979Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1467566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1468142Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1468724Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 19%] 2025-07-17T09:17:08.1469194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0066s] [ 19%] 2025-07-17T09:17:08.1469664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0027s] [ 19%] 2025-07-17T09:17:08.1470125Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 19%] 2025-07-17T09:17:08.1470706Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1471240Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0027s] [ 19%] 2025-07-17T09:17:08.1471755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0027s] [ 19%] 2025-07-17T09:17:08.1472224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1472685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 19%] 2025-07-17T09:17:08.1473150Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0050s] [ 19%] 2025-07-17T09:17:08.1473612Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 19%] 2025-07-17T09:17:08.1474133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1474593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1475054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0027s] [ 19%] 2025-07-17T09:17:08.1475513Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 19%] 2025-07-17T09:17:08.1475975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1476428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1476897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 19%] 2025-07-17T09:17:08.1477357Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1477949Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1478472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 19%] 2025-07-17T09:17:08.1479008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1479472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1479943Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 19%] 2025-07-17T09:17:08.1480408Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1480869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1481331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 19%] 2025-07-17T09:17:08.1481788Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1482248Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1482713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1483180Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1483640Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1484096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1484560Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1485133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1485652Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1486159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1486633Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1487104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 19%] 2025-07-17T09:17:08.1487562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 19%] 2025-07-17T09:17:08.1488028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1488483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 20%] 2025-07-17T09:17:08.1488948Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1489401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1489858Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1490325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 20%] 2025-07-17T09:17:08.1490784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 20%] 2025-07-17T09:17:08.1491242Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1491694Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1492431Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1493068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1493702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1494288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1494867Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1495453Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1496038Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1496622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1497203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1497781Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1498366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1498936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1499640Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1500271Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1500892Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1501481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1502065Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1502654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1503241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1503820Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1504407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1504993Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1505576Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1506170Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1506876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1507508Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1508127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1508718Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1509297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1509884Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1510461Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1511037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1511621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1512202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1512792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1513371Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1514124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1514827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1515466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1516050Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1516634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1517211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1517791Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1518362Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1518943Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1519521Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1520097Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1520679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 20%] 2025-07-17T09:17:08.1521211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0074s] [ 20%] 2025-07-17T09:17:08.1521775Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 20%] 2025-07-17T09:17:08.1522292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1522755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1523223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0028s] [ 20%] 2025-07-17T09:17:08.1523679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 20%] 2025-07-17T09:17:08.1524141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1524593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1525054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0058s] [ 20%] 2025-07-17T09:17:08.1525503Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 20%] 2025-07-17T09:17:08.1525966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1526414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1526869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0028s] [ 20%] 2025-07-17T09:17:08.1527323Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 20%] 2025-07-17T09:17:08.1527773Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1528232Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1528846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0061s] [ 20%] 2025-07-17T09:17:08.1529355Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1529817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 20%] 2025-07-17T09:17:08.1530271Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1530742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1531200Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1531660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1532118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1532575Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0060s] [ 20%] 2025-07-17T09:17:08.1533030Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1533478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 20%] 2025-07-17T09:17:08.1533945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1534401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1534858Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1535309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1535937Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1536452Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1536907Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1537369Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1537829Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1538297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1538755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1539219Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1539683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1540137Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1540590Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1541043Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1541504Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1541957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1542410Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1542980Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1543490Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1544016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0070s] [ 20%] 2025-07-17T09:17:08.1544475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 20%] 2025-07-17T09:17:08.1544932Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1545380Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1545840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 20%] 2025-07-17T09:17:08.1546302Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 20%] 2025-07-17T09:17:08.1546757Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1547217Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 20%] 2025-07-17T09:17:08.1547664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0067s] [ 20%] 2025-07-17T09:17:08.1548115Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 20%] 2025-07-17T09:17:08.1548563Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 20%] 2025-07-17T09:17:08.1549016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1549471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 21%] 2025-07-17T09:17:08.1550033Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 21%] 2025-07-17T09:17:08.1550538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1551040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1551500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0067s] [ 21%] 2025-07-17T09:17:08.1551961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 21%] 2025-07-17T09:17:08.1552419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1552871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1553333Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1553796Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1554330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1554795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1555248Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0067s] [ 21%] 2025-07-17T09:17:08.1555702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 21%] 2025-07-17T09:17:08.1556155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1556607Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1557196Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1557705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1558225Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1558678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1559140Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1559596Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 21%] 2025-07-17T09:17:08.1560058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1560512Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1560973Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1561433Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1561885Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1562348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1562801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 21%] 2025-07-17T09:17:08.1563258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1563708Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1564274Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1564785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1565285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1565740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1566198Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1566663Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0068s] [ 21%] 2025-07-17T09:17:08.1567127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0027s] [ 21%] 2025-07-17T09:17:08.1567586Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 21%] 2025-07-17T09:17:08.1568046Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 21%] 2025-07-17T09:17:08.1568510Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0027s] [ 21%] 2025-07-17T09:17:08.1568974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0027s] [ 21%] 2025-07-17T09:17:08.1569435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0028s] [ 21%] 2025-07-17T09:17:08.1569897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 21%] 2025-07-17T09:17:08.1570354Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0076s] [ 21%] 2025-07-17T09:17:08.1570808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0056s] [ 21%] 2025-07-17T09:17:08.1571372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 21%] 2025-07-17T09:17:08.1571891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0039s] [ 21%] 2025-07-17T09:17:08.1572526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 21%] 2025-07-17T09:17:08.1572979Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 21%] 2025-07-17T09:17:08.1573441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0052s] [ 21%] 2025-07-17T09:17:08.1573895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0045s] [ 21%] 2025-07-17T09:17:08.1574356Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0081s] [ 21%] 2025-07-17T09:17:08.1574815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0056s] [ 21%] 2025-07-17T09:17:08.1575277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0067s] [ 21%] 2025-07-17T09:17:08.1575739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0053s] [ 21%] 2025-07-17T09:17:08.1576200Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 21%] 2025-07-17T09:17:08.1576660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 21%] 2025-07-17T09:17:08.1577124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 21%] 2025-07-17T09:17:08.1577584Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1578041Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0068s] [ 21%] 2025-07-17T09:17:08.1578601Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 21%] 2025-07-17T09:17:08.1579109Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 21%] 2025-07-17T09:17:08.1579616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1580072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 21%] 2025-07-17T09:17:08.1580537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1580989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 21%] 2025-07-17T09:17:08.1581447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 21%] 2025-07-17T09:17:08.1581911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 21%] 2025-07-17T09:17:08.1582379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1582834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1583291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0043s] [ 21%] 2025-07-17T09:17:08.1583758Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1584221Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1584681Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1585139Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1585721Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1586224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1586731Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 21%] 2025-07-17T09:17:08.1587187Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 21%] 2025-07-17T09:17:08.1587649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1588111Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 21%] 2025-07-17T09:17:08.1588564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 21%] 2025-07-17T09:17:08.1589020Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 21%] 2025-07-17T09:17:08.1589478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0082s] [ 21%] 2025-07-17T09:17:08.1589935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1590386Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 21%] 2025-07-17T09:17:08.1590840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 21%] 2025-07-17T09:17:08.1591301Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 21%] 2025-07-17T09:17:08.1591756Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1592212Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1592775Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1593281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0072s] [ 21%] 2025-07-17T09:17:08.1593780Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 21%] 2025-07-17T09:17:08.1594342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1594802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 21%] 2025-07-17T09:17:08.1595250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 21%] 2025-07-17T09:17:08.1595711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1596156Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1596606Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1597061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0061s] [ 21%] 2025-07-17T09:17:08.1597517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1597973Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 21%] 2025-07-17T09:17:08.1598426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 21%] 2025-07-17T09:17:08.1598892Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1599345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 21%] 2025-07-17T09:17:08.1599869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 21%] 2025-07-17T09:17:08.1600462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 21%] 2025-07-17T09:17:08.1600978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0061s] [ 21%] 2025-07-17T09:17:08.1601432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1601886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1602337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 21%] 2025-07-17T09:17:08.1602793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1603249Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1603702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 21%] 2025-07-17T09:17:08.1604156Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1604623Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1605080Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 22%] 2025-07-17T09:17:08.1605540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1605991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 22%] 2025-07-17T09:17:08.1606455Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 22%] 2025-07-17T09:17:08.1606968Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1607526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 22%] 2025-07-17T09:17:08.1608039Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1608492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1608950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 22%] 2025-07-17T09:17:08.1609399Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1609852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1610301Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1610761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1611216Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 22%] 2025-07-17T09:17:08.1611665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 22%] 2025-07-17T09:17:08.1612120Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0054s] [ 22%] 2025-07-17T09:17:08.1612575Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 22%] 2025-07-17T09:17:08.1613035Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 22%] 2025-07-17T09:17:08.1613486Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 22%] 2025-07-17T09:17:08.1613943Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0027s] [ 22%] 2025-07-17T09:17:08.1614554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 22%] 2025-07-17T09:17:08.1615051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 22%] 2025-07-17T09:17:08.1615506Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 22%] 2025-07-17T09:17:08.1615966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0051s] [ 22%] 2025-07-17T09:17:08.1616425Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1616876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1617321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0049s] [ 22%] 2025-07-17T09:17:08.1617772Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1618225Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 22%] 2025-07-17T09:17:08.1618676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1619118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1619587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 22%] 2025-07-17T09:17:08.1620042Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1620501Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 22%] 2025-07-17T09:17:08.1620948Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1621519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1622037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1622539Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 22%] 2025-07-17T09:17:08.1622995Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1623450Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 22%] 2025-07-17T09:17:08.1623912Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1624356Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1624813Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0041s] [ 22%] 2025-07-17T09:17:08.1625268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0044s] [ 22%] 2025-07-17T09:17:08.1625717Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1626168Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 22%] 2025-07-17T09:17:08.1626616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1627079Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1627532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1627986Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1628567Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1629075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1629583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1630038Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1630496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1630949Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1631404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1631854Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 22%] 2025-07-17T09:17:08.1632303Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1632761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1633210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1633664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 22%] 2025-07-17T09:17:08.1634190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1634653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0066s] [ 22%] 2025-07-17T09:17:08.1635105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 22%] 2025-07-17T09:17:08.1635694Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1636212Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1636727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 22%] 2025-07-17T09:17:08.1637185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 22%] 2025-07-17T09:17:08.1637642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1638098Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1638551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0057s] [ 22%] 2025-07-17T09:17:08.1638998Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1639448Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1639889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1640341Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1640789Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 22%] 2025-07-17T09:17:08.1641237Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1641681Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1642132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0061s] [ 22%] 2025-07-17T09:17:08.1642714Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1643225Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1643728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 22%] 2025-07-17T09:17:08.1644182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1644641Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1645091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1645547Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1645994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0060s] [ 22%] 2025-07-17T09:17:08.1646444Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 22%] 2025-07-17T09:17:08.1646889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1647332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 22%] 2025-07-17T09:17:08.1647782Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1648227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1648678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 22%] 2025-07-17T09:17:08.1649126Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1649647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1650211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1650727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1651176Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 22%] 2025-07-17T09:17:08.1651634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1652090Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1652543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1652990Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 22%] 2025-07-17T09:17:08.1653440Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 22%] 2025-07-17T09:17:08.1653891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1654344Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1654784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1655237Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 22%] 2025-07-17T09:17:08.1655686Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1656129Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1656578Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 22%] 2025-07-17T09:17:08.1657149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 22%] 2025-07-17T09:17:08.1657669Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 22%] 2025-07-17T09:17:08.1658182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 22%] 2025-07-17T09:17:08.1658645Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1659117Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 23%] 2025-07-17T09:17:08.1659582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 23%] 2025-07-17T09:17:08.1660043Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1660505Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1660965Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 23%] 2025-07-17T09:17:08.1661422Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 23%] 2025-07-17T09:17:08.1661890Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1662345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1662805Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 23%] 2025-07-17T09:17:08.1663264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 23%] 2025-07-17T09:17:08.1663718Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1664295Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 23%] 2025-07-17T09:17:08.1664814Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 23%] 2025-07-17T09:17:08.1665337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1665804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 23%] 2025-07-17T09:17:08.1666266Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 23%] 2025-07-17T09:17:08.1666743Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1667206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1667678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 23%] 2025-07-17T09:17:08.1668140Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 23%] 2025-07-17T09:17:08.1668601Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 23%] 2025-07-17T09:17:08.1669061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1669524Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 23%] 2025-07-17T09:17:08.1669985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 23%] 2025-07-17T09:17:08.1670444Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1670905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 23%] 2025-07-17T09:17:08.1671516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 23%] 2025-07-17T09:17:08.1672030Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 23%] 2025-07-17T09:17:08.1672494Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 23%] 2025-07-17T09:17:08.1672969Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1673437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 23%] 2025-07-17T09:17:08.1673963Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 23%] 2025-07-17T09:17:08.1674444Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1674909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1675380Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 23%] 2025-07-17T09:17:08.1675846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 23%] 2025-07-17T09:17:08.1676313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 23%] 2025-07-17T09:17:08.1676779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1677234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 23%] 2025-07-17T09:17:08.1677693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1678220Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1678820Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1679333Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0047s] [ 23%] 2025-07-17T09:17:08.1679795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 23%] 2025-07-17T09:17:08.1680397Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1680982Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1681570Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1682149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1682740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1683329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1683906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1684501Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1685083Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1685713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1686540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1687162Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1687748Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1688326Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1688906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1689483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1690068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1690657Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1691242Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1691832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1692426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1693063Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1693753Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1694379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1694966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1695553Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1696130Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1696709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1697289Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1697871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1698448Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1699024Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1699613Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1700252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1700927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1701559Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1702147Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1702742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1703329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1703908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1704488Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1705066Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1705646Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1706223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1706802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1707445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1708134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1708768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 23%] 2025-07-17T09:17:08.1709238Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 23%] 2025-07-17T09:17:08.1709704Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 23%] 2025-07-17T09:17:08.1710169Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1710624Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 23%] 2025-07-17T09:17:08.1711093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 23%] 2025-07-17T09:17:08.1711558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 23%] 2025-07-17T09:17:08.1712024Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 23%] 2025-07-17T09:17:08.1712494Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 23%] 2025-07-17T09:17:08.1712951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 23%] 2025-07-17T09:17:08.1713411Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 23%] 2025-07-17T09:17:08.1713865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 23%] 2025-07-17T09:17:08.1714389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 23%] 2025-07-17T09:17:08.1714979Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 23%] 2025-07-17T09:17:08.1715506Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 23%] 2025-07-17T09:17:08.1716026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 23%] 2025-07-17T09:17:08.1716481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 23%] 2025-07-17T09:17:08.1716955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 23%] 2025-07-17T09:17:08.1717421Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 23%] 2025-07-17T09:17:08.1717886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 23%] 2025-07-17T09:17:08.1718345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 23%] 2025-07-17T09:17:08.1718818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 23%] 2025-07-17T09:17:08.1719285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 23%] 2025-07-17T09:17:08.1719745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 23%] 2025-07-17T09:17:08.1720220Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 24%] 2025-07-17T09:17:08.1720679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 24%] 2025-07-17T09:17:08.1721145Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 24%] 2025-07-17T09:17:08.1721599Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 24%] 2025-07-17T09:17:08.1722167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 24%] 2025-07-17T09:17:08.1722682Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 24%] 2025-07-17T09:17:08.1723192Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 24%] 2025-07-17T09:17:08.1723659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 24%] 2025-07-17T09:17:08.1724116Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 24%] 2025-07-17T09:17:08.1724584Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 24%] 2025-07-17T09:17:08.1725049Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 24%] 2025-07-17T09:17:08.1725517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 24%] 2025-07-17T09:17:08.1725983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 24%] 2025-07-17T09:17:08.1726450Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 24%] 2025-07-17T09:17:08.1726917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 24%] 2025-07-17T09:17:08.1727378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 24%] 2025-07-17T09:17:08.1727841Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 24%] 2025-07-17T09:17:08.1728300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 24%] 2025-07-17T09:17:08.1728764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 24%] 2025-07-17T09:17:08.1729323Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 24%] 2025-07-17T09:17:08.1729847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 24%] 2025-07-17T09:17:08.1730367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 24%] 2025-07-17T09:17:08.1730827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 24%] 2025-07-17T09:17:08.1731302Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 24%] 2025-07-17T09:17:08.1731760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 24%] 2025-07-17T09:17:08.1732351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1732941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1733529Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1734110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1734695Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1735284Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1735869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1736588Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1737224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1737850Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1738433Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1739009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1739589Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1740170Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1740741Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1741324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1741916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1742498Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1743082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1743769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1744412Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1745046Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1745632Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1746213Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1746793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1747377Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1747962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1748535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1749121Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1749702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1750280Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1750966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1751606Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1752247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1752838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1753424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1754076Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1754662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1755247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1755828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1756413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1757008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1757583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1758285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1758927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1759593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1760174Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1760755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 24%] 2025-07-17T09:17:08.1761225Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0050s] [ 24%] 2025-07-17T09:17:08.1761682Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1762145Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 24%] 2025-07-17T09:17:08.1762603Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 24%] 2025-07-17T09:17:08.1763063Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1763527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1763987Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1764458Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 24%] 2025-07-17T09:17:08.1764912Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 24%] 2025-07-17T09:17:08.1765481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1765988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1766488Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1766946Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0027s] [ 24%] 2025-07-17T09:17:08.1767403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 24%] 2025-07-17T09:17:08.1767862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1768317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1768781Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 24%] 2025-07-17T09:17:08.1769248Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1769702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1770163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1770623Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1771089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1771550Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1772014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1772572Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1773086Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 24%] 2025-07-17T09:17:08.1773591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1774042Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1774508Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1774962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1775421Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1775872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1776338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1776808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 24%] 2025-07-17T09:17:08.1777263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1777725Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1778187Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0028s] [ 24%] 2025-07-17T09:17:08.1778652Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 24%] 2025-07-17T09:17:08.1779108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1779685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1780201Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1780701Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 24%] 2025-07-17T09:17:08.1781160Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 24%] 2025-07-17T09:17:08.1781615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 25%] 2025-07-17T09:17:08.1782075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 25%] 2025-07-17T09:17:08.1782526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 25%] 2025-07-17T09:17:08.1782982Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 25%] 2025-07-17T09:17:08.1783442Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 25%] 2025-07-17T09:17:08.1784024Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1784609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1785183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1785765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1786348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1787037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1787692Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1788314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1788893Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1789473Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1790042Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1790618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1791189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1791767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1792339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1792907Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1793492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1794253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1794898Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1795537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1796124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1796709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1797281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1797868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1798448Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1799022Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1799596Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1800165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1800744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1801373Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1802127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1802765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1803354Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1803941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1804520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1805093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1805683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1806263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1806849Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1807429Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1808002Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1808637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1809307Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1809929Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1810515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1811092Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1811670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1812243Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1812725Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0044s] [ 25%] 2025-07-17T09:17:08.1813199Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 25%] 2025-07-17T09:17:08.1813660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 25%] 2025-07-17T09:17:08.1814134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 25%] 2025-07-17T09:17:08.1814610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 25%] 2025-07-17T09:17:08.1815082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 25%] 2025-07-17T09:17:08.1815546Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 25%] 2025-07-17T09:17:08.1816139Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 25%] 2025-07-17T09:17:08.1816656Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 25%] 2025-07-17T09:17:08.1817163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 25%] 2025-07-17T09:17:08.1817627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 25%] 2025-07-17T09:17:08.1818093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 25%] 2025-07-17T09:17:08.1818567Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 25%] 2025-07-17T09:17:08.1819027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 25%] 2025-07-17T09:17:08.1819492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 25%] 2025-07-17T09:17:08.1819954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 25%] 2025-07-17T09:17:08.1820421Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0047s] [ 25%] 2025-07-17T09:17:08.1820888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 25%] 2025-07-17T09:17:08.1821352Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 25%] 2025-07-17T09:17:08.1821837Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 25%] 2025-07-17T09:17:08.1822311Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 25%] 2025-07-17T09:17:08.1822784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 25%] 2025-07-17T09:17:08.1823378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 25%] 2025-07-17T09:17:08.1823900Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 25%] 2025-07-17T09:17:08.1824423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0043s] [ 25%] 2025-07-17T09:17:08.1824885Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 25%] 2025-07-17T09:17:08.1825356Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 25%] 2025-07-17T09:17:08.1825818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 25%] 2025-07-17T09:17:08.1826287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 25%] 2025-07-17T09:17:08.1826755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 25%] 2025-07-17T09:17:08.1827218Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 25%] 2025-07-17T09:17:08.1827682Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 25%] 2025-07-17T09:17:08.1828149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 25%] 2025-07-17T09:17:08.1828621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 25%] 2025-07-17T09:17:08.1829081Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 25%] 2025-07-17T09:17:08.1829552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 25%] 2025-07-17T09:17:08.1830091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 25%] 2025-07-17T09:17:08.1830662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 25%] 2025-07-17T09:17:08.1831180Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 25%] 2025-07-17T09:17:08.1831644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 25%] 2025-07-17T09:17:08.1832114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 25%] 2025-07-17T09:17:08.1832572Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 25%] 2025-07-17T09:17:08.1833036Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 25%] 2025-07-17T09:17:08.1833539Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 25%] 2025-07-17T09:17:08.1834158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 25%] 2025-07-17T09:17:08.1834688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0040s] [ 25%] 2025-07-17T09:17:08.1835152Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0047s] [ 25%] 2025-07-17T09:17:08.1835615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0044s] [ 25%] 2025-07-17T09:17:08.1836214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1836815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1837409Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1838203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1838857Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1839447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1840040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0017s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1840629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1841210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1841795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1842374Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1842948Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1843538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1844121Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1844705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 25%] 2025-07-17T09:17:08.1845464Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1846099Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1846689Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1847276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1847868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1848461Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1849053Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1849646Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1850227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1850817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1851400Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1851975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1852716Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1853349Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1853937Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1854524Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1855101Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1855689Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1856277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1856868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1857457Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1858045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1858634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1859214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1859962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1860606Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1861190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1861778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1862355Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1862944Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1863533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1864120Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1864704Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1865177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0072s] [ 26%] 2025-07-17T09:17:08.1865653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 26%] 2025-07-17T09:17:08.1866128Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 26%] 2025-07-17T09:17:08.1866644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 26%] 2025-07-17T09:17:08.1867219Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 26%] 2025-07-17T09:17:08.1867747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 26%] 2025-07-17T09:17:08.1868223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 26%] 2025-07-17T09:17:08.1868694Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 26%] 2025-07-17T09:17:08.1869167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0049s] [ 26%] 2025-07-17T09:17:08.1869636Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 26%] 2025-07-17T09:17:08.1870098Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 26%] 2025-07-17T09:17:08.1870566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 26%] 2025-07-17T09:17:08.1871031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 26%] 2025-07-17T09:17:08.1871498Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 26%] 2025-07-17T09:17:08.1871955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 26%] 2025-07-17T09:17:08.1872421Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 26%] 2025-07-17T09:17:08.1872895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0053s] [ 26%] 2025-07-17T09:17:08.1873360Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 26%] 2025-07-17T09:17:08.1874004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 26%] 2025-07-17T09:17:08.1874532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 26%] 2025-07-17T09:17:08.1875068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 26%] 2025-07-17T09:17:08.1875537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 26%] 2025-07-17T09:17:08.1876017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 26%] 2025-07-17T09:17:08.1876489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 26%] 2025-07-17T09:17:08.1876949Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0052s] [ 26%] 2025-07-17T09:17:08.1877412Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 26%] 2025-07-17T09:17:08.1877869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 26%] 2025-07-17T09:17:08.1878335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 26%] 2025-07-17T09:17:08.1878796Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 26%] 2025-07-17T09:17:08.1879263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 26%] 2025-07-17T09:17:08.1879731Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 26%] 2025-07-17T09:17:08.1880192Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 26%] 2025-07-17T09:17:08.1880664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 26%] 2025-07-17T09:17:08.1881255Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 26%] 2025-07-17T09:17:08.1881781Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 26%] 2025-07-17T09:17:08.1882292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 26%] 2025-07-17T09:17:08.1882772Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 26%] 2025-07-17T09:17:08.1883251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 26%] 2025-07-17T09:17:08.1883719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 26%] 2025-07-17T09:17:08.1884190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 26%] 2025-07-17T09:17:08.1884651Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 26%] 2025-07-17T09:17:08.1885118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 26%] 2025-07-17T09:17:08.1885573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 26%] 2025-07-17T09:17:08.1886034Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 26%] 2025-07-17T09:17:08.1886508Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 26%] 2025-07-17T09:17:08.1886970Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 26%] 2025-07-17T09:17:08.1887442Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 26%] 2025-07-17T09:17:08.1887903Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 26%] 2025-07-17T09:17:08.1888610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1889251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1889890Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1890481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1891091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1891688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1892274Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1892870Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1893455Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1894035Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1894616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1895197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1895895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1896548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1897174Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1897759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1898354Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1898942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1899530Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1900113Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1900699Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1901289Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1901871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1902484Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1903186Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1903831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1904459Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1905038Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1905628Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1906209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1906794Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1907376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1907962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1908553Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1909136Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 26%] 2025-07-17T09:17:08.1909727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1910429Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1911069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1911703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1912291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1912880Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1913463Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1914111Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1914695Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1915275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1915866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1916450Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1917028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1917711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0052s] [ 27%] 2025-07-17T09:17:08.1918234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 27%] 2025-07-17T09:17:08.1918761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 27%] 2025-07-17T09:17:08.1919231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 27%] 2025-07-17T09:17:08.1919705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0028s] [ 27%] 2025-07-17T09:17:08.1920174Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 27%] 2025-07-17T09:17:08.1920632Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 27%] 2025-07-17T09:17:08.1921100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 27%] 2025-07-17T09:17:08.1921559Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 27%] 2025-07-17T09:17:08.1922022Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 27%] 2025-07-17T09:17:08.1922481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 27%] 2025-07-17T09:17:08.1922936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1923401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0028s] [ 27%] 2025-07-17T09:17:08.1923858Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 27%] 2025-07-17T09:17:08.1924319Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 27%] 2025-07-17T09:17:08.1924898Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 27%] 2025-07-17T09:17:08.1925424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 27%] 2025-07-17T09:17:08.1925956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 27%] 2025-07-17T09:17:08.1926417Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 27%] 2025-07-17T09:17:08.1926889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 27%] 2025-07-17T09:17:08.1927359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1927829Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1928291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 27%] 2025-07-17T09:17:08.1928761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 27%] 2025-07-17T09:17:08.1929228Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1929687Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1930152Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 27%] 2025-07-17T09:17:08.1930610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 27%] 2025-07-17T09:17:08.1931077Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1931533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1932106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 27%] 2025-07-17T09:17:08.1932615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 27%] 2025-07-17T09:17:08.1933139Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1933661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1934143Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 27%] 2025-07-17T09:17:08.1934616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 27%] 2025-07-17T09:17:08.1935079Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1935549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 27%] 2025-07-17T09:17:08.1936011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 27%] 2025-07-17T09:17:08.1936481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 27%] 2025-07-17T09:17:08.1936945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 27%] 2025-07-17T09:17:08.1937403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1937864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 27%] 2025-07-17T09:17:08.1938316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 27%] 2025-07-17T09:17:08.1938782Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1939353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 27%] 2025-07-17T09:17:08.1939873Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 27%] 2025-07-17T09:17:08.1940386Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 27%] 2025-07-17T09:17:08.1940978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1941567Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1942146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1942728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1943318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1943899Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1944481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1945062Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1945646Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1946280Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1946952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1947580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1948159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1948754Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1949336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1949908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1950497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1951077Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1951661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1952247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1952832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1953494Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1954264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1954916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1955509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1956091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1956674Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1957243Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1957833Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1958419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1958993Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1959574Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1960163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1960812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1961507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1962131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1962722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1963315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1963894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1964482Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1965061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1965642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1966222Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1966801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1967383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1967959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1968696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1969322Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 27%] 2025-07-17T09:17:08.1969796Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 27%] 2025-07-17T09:17:08.1970275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0054s] [ 27%] 2025-07-17T09:17:08.1970740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0153s] [ 27%] 2025-07-17T09:17:08.1971211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0100s] [ 27%] 2025-07-17T09:17:08.1971687Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0055s] [ 27%] 2025-07-17T09:17:08.1972167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0054s] [ 27%] 2025-07-17T09:17:08.1972647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0153s] [ 27%] 2025-07-17T09:17:08.1973116Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0101s] [ 28%] 2025-07-17T09:17:08.1973589Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0056s] [ 28%] 2025-07-17T09:17:08.1974051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0056s] [ 28%] 2025-07-17T09:17:08.1974514Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0155s] [ 28%] 2025-07-17T09:17:08.1974973Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0102s] [ 28%] 2025-07-17T09:17:08.1975562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0058s] [ 28%] 2025-07-17T09:17:08.1976088Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0056s] [ 28%] 2025-07-17T09:17:08.1976611Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0155s] [ 28%] 2025-07-17T09:17:08.1977078Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0100s] [ 28%] 2025-07-17T09:17:08.1977551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0059s] [ 28%] 2025-07-17T09:17:08.1978024Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0059s] [ 28%] 2025-07-17T09:17:08.1978488Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0171s] [ 28%] 2025-07-17T09:17:08.1978965Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0112s] [ 28%] 2025-07-17T09:17:08.1979449Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 28%] 2025-07-17T09:17:08.1979918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0059s] [ 28%] 2025-07-17T09:17:08.1980394Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0170s] [ 28%] 2025-07-17T09:17:08.1980864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0111s] [ 28%] 2025-07-17T09:17:08.1981336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0059s] [ 28%] 2025-07-17T09:17:08.1981798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0059s] [ 28%] 2025-07-17T09:17:08.1982272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0170s] [ 28%] 2025-07-17T09:17:08.1982872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0109s] [ 28%] 2025-07-17T09:17:08.1983395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0059s] [ 28%] 2025-07-17T09:17:08.1983912Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0058s] [ 28%] 2025-07-17T09:17:08.1984378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0170s] [ 28%] 2025-07-17T09:17:08.1984848Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0110s] [ 28%] 2025-07-17T09:17:08.1985316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0058s] [ 28%] 2025-07-17T09:17:08.1985789Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0058s] [ 28%] 2025-07-17T09:17:08.2005481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0170s] [ 28%] 2025-07-17T09:17:08.2006044Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0110s] [ 28%] 2025-07-17T09:17:08.2006540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0059s] [ 28%] 2025-07-17T09:17:08.2007014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0058s] [ 28%] 2025-07-17T09:17:08.2007499Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0171s] [ 28%] 2025-07-17T09:17:08.2007972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0110s] [ 28%] 2025-07-17T09:17:08.2008440Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0058s] [ 28%] 2025-07-17T09:17:08.2009052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0058s] [ 28%] 2025-07-17T09:17:08.2009667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0171s] [ 28%] 2025-07-17T09:17:08.2010194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0110s] [ 28%] 2025-07-17T09:17:08.2010664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0059s] [ 28%] 2025-07-17T09:17:08.2011146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0058s] [ 28%] 2025-07-17T09:17:08.2011619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0171s] [ 28%] 2025-07-17T09:17:08.2012080Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0110s] [ 28%] 2025-07-17T09:17:08.2012689Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2013282Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2013887Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2014500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2015095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2015696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2016281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2017025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2017661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2018252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2018839Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2019417Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2020007Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2020595Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2021176Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2021762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2022353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2022945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2023531Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2024297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2024941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2025527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2026122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2026712Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2027291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2027882Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2028460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2029044Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2029640Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2030227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2030815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2031564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2032218Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2032811Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2033406Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2034088Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2034692Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2035290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2035879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2036463Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2037057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2037641Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2038294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2039025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2039663Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2040250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2040833Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2041424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 28%] 2025-07-17T09:17:08.2041904Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0058s] [ 28%] 2025-07-17T09:17:08.2042376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0053s] [ 28%] 2025-07-17T09:17:08.2042850Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0172s] [ 28%] 2025-07-17T09:17:08.2043315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0116s] [ 28%] 2025-07-17T09:17:08.2043793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0054s] [ 28%] 2025-07-17T09:17:08.2044272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0053s] [ 28%] 2025-07-17T09:17:08.2044740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0172s] [ 28%] 2025-07-17T09:17:08.2045213Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0115s] [ 28%] 2025-07-17T09:17:08.2045793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0053s] [ 28%] 2025-07-17T09:17:08.2046321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0053s] [ 28%] 2025-07-17T09:17:08.2046832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0171s] [ 28%] 2025-07-17T09:17:08.2047298Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0115s] [ 28%] 2025-07-17T09:17:08.2047774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0054s] [ 28%] 2025-07-17T09:17:08.2048249Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0053s] [ 28%] 2025-07-17T09:17:08.2048721Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0172s] [ 28%] 2025-07-17T09:17:08.2049182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0113s] [ 28%] 2025-07-17T09:17:08.2049659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 28%] 2025-07-17T09:17:08.2050127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0062s] [ 28%] 2025-07-17T09:17:08.2050597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0191s] [ 28%] 2025-07-17T09:17:08.2051069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0128s] [ 28%] 2025-07-17T09:17:08.2051542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 28%] 2025-07-17T09:17:08.2052018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0059s] [ 28%] 2025-07-17T09:17:08.2052483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0192s] [ 28%] 2025-07-17T09:17:08.2053065Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0128s] [ 28%] 2025-07-17T09:17:08.2053585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0060s] [ 28%] 2025-07-17T09:17:08.2054102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0059s] [ 28%] 2025-07-17T09:17:08.2054569Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0191s] [ 28%] 2025-07-17T09:17:08.2055039Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0125s] [ 29%] 2025-07-17T09:17:08.2055509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0060s] [ 29%] 2025-07-17T09:17:08.2055970Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0059s] [ 29%] 2025-07-17T09:17:08.2056439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0191s] [ 29%] 2025-07-17T09:17:08.2056902Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0127s] [ 29%] 2025-07-17T09:17:08.2057375Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 29%] 2025-07-17T09:17:08.2057846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0059s] [ 29%] 2025-07-17T09:17:08.2058310Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0191s] [ 29%] 2025-07-17T09:17:08.2058781Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0127s] [ 29%] 2025-07-17T09:17:08.2059256Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 29%] 2025-07-17T09:17:08.2059734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0059s] [ 29%] 2025-07-17T09:17:08.2060314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0191s] [ 29%] 2025-07-17T09:17:08.2060846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0126s] [ 29%] 2025-07-17T09:17:08.2061360Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0060s] [ 29%] 2025-07-17T09:17:08.2061824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0059s] [ 29%] 2025-07-17T09:17:08.2062293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0190s] [ 29%] 2025-07-17T09:17:08.2062752Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0127s] [ 29%] 2025-07-17T09:17:08.2063225Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0059s] [ 29%] 2025-07-17T09:17:08.2063692Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0059s] [ 29%] 2025-07-17T09:17:08.2064164Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0190s] [ 29%] 2025-07-17T09:17:08.2064629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0126s] [ 29%] 2025-07-17T09:17:08.2065224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2065832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2066426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2067097Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2067824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2068464Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2069056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2069644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2070232Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2070815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2071394Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2071981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2072574Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2073166Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2073762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2074544Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2075260Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2075909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2076507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2077095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2077686Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2078284Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2078872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2079464Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2080068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2080649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2081236Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2081864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2082559Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2083193Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2083781Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2084368Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2084955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2085552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2086149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2086733Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2087335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2087930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2088524Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2089181Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2089862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2090522Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2091109Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2091692Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2092284Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2092867Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2093459Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2094044Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 29%] 2025-07-17T09:17:08.2094516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 29%] 2025-07-17T09:17:08.2094993Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 29%] 2025-07-17T09:17:08.2095458Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 29%] 2025-07-17T09:17:08.2095921Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 29%] 2025-07-17T09:17:08.2096443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 29%] 2025-07-17T09:17:08.2097011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 29%] 2025-07-17T09:17:08.2097528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 29%] 2025-07-17T09:17:08.2097992Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 29%] 2025-07-17T09:17:08.2098470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 29%] 2025-07-17T09:17:08.2098947Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 29%] 2025-07-17T09:17:08.2099418Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 29%] 2025-07-17T09:17:08.2099874Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 29%] 2025-07-17T09:17:08.2100347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 29%] 2025-07-17T09:17:08.2100814Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 29%] 2025-07-17T09:17:08.2101276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 29%] 2025-07-17T09:17:08.2101740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 29%] 2025-07-17T09:17:08.2102217Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 29%] 2025-07-17T09:17:08.2102692Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 29%] 2025-07-17T09:17:08.2103155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0059s] [ 29%] 2025-07-17T09:17:08.2103731Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0044s] [ 29%] 2025-07-17T09:17:08.2104264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 29%] 2025-07-17T09:17:08.2104778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 29%] 2025-07-17T09:17:08.2105247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0059s] [ 29%] 2025-07-17T09:17:08.2105710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0044s] [ 29%] 2025-07-17T09:17:08.2106179Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 29%] 2025-07-17T09:17:08.2106651Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 29%] 2025-07-17T09:17:08.2107121Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0059s] [ 29%] 2025-07-17T09:17:08.2107585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0044s] [ 29%] 2025-07-17T09:17:08.2108047Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 29%] 2025-07-17T09:17:08.2108511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 29%] 2025-07-17T09:17:08.2108970Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0059s] [ 29%] 2025-07-17T09:17:08.2109434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0044s] [ 29%] 2025-07-17T09:17:08.2109905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 29%] 2025-07-17T09:17:08.2110377Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 29%] 2025-07-17T09:17:08.2110960Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 29%] 2025-07-17T09:17:08.2111485Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0045s] [ 29%] 2025-07-17T09:17:08.2112010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 29%] 2025-07-17T09:17:08.2112479Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 29%] 2025-07-17T09:17:08.2112953Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0059s] [ 29%] 2025-07-17T09:17:08.2113415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0044s] [ 29%] 2025-07-17T09:17:08.2113951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 29%] 2025-07-17T09:17:08.2114424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 29%] 2025-07-17T09:17:08.2114883Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0060s] [ 29%] 2025-07-17T09:17:08.2115346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0044s] [ 29%] 2025-07-17T09:17:08.2115814Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 29%] 2025-07-17T09:17:08.2116287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 29%] 2025-07-17T09:17:08.2116744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0059s] [ 29%] 2025-07-17T09:17:08.2117206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0044s] [ 30%] 2025-07-17T09:17:08.2117803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2118634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2119279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2119857Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2120451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2121040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2121621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2122214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2122795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2123377Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2123962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2124533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2125114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2125841Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2126477Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2127062Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2127676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2128267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2128848Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2129438Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2130031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2130616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2131203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2131784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2132366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2133103Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2133730Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2134309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2134893Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2135480Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2136058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2136636Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2137229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2137812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2138400Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2138983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2139570Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2140281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2140916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2141562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2142149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2142737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2143321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2143893Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2144478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2145068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2145646Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2146228Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2146699Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 30%] 2025-07-17T09:17:08.2147295Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 30%] 2025-07-17T09:17:08.2147814Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 30%] 2025-07-17T09:17:08.2148326Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 30%] 2025-07-17T09:17:08.2148802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 30%] 2025-07-17T09:17:08.2149268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 30%] 2025-07-17T09:17:08.2149739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 30%] 2025-07-17T09:17:08.2150198Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 30%] 2025-07-17T09:17:08.2150662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 30%] 2025-07-17T09:17:08.2151125Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 30%] 2025-07-17T09:17:08.2151581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 30%] 2025-07-17T09:17:08.2152039Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 30%] 2025-07-17T09:17:08.2152505Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 30%] 2025-07-17T09:17:08.2152975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 30%] 2025-07-17T09:17:08.2153431Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 30%] 2025-07-17T09:17:08.2153966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 30%] 2025-07-17T09:17:08.2154566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 30%] 2025-07-17T09:17:08.2155089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 30%] 2025-07-17T09:17:08.2155615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 30%] 2025-07-17T09:17:08.2156075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 30%] 2025-07-17T09:17:08.2156557Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 30%] 2025-07-17T09:17:08.2157018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 30%] 2025-07-17T09:17:08.2157486Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 30%] 2025-07-17T09:17:08.2157951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 30%] 2025-07-17T09:17:08.2158410Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 30%] 2025-07-17T09:17:08.2158872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 30%] 2025-07-17T09:17:08.2159325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 30%] 2025-07-17T09:17:08.2159788Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 30%] 2025-07-17T09:17:08.2160252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 30%] 2025-07-17T09:17:08.2160717Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 30%] 2025-07-17T09:17:08.2161173Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 30%] 2025-07-17T09:17:08.2161741Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 30%] 2025-07-17T09:17:08.2162263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 30%] 2025-07-17T09:17:08.2162777Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 30%] 2025-07-17T09:17:08.2163247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 30%] 2025-07-17T09:17:08.2163711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 30%] 2025-07-17T09:17:08.2164185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 30%] 2025-07-17T09:17:08.2164664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 30%] 2025-07-17T09:17:08.2165127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 30%] 2025-07-17T09:17:08.2165590Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 30%] 2025-07-17T09:17:08.2166053Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 30%] 2025-07-17T09:17:08.2166515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 30%] 2025-07-17T09:17:08.2166975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 30%] 2025-07-17T09:17:08.2167436Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 30%] 2025-07-17T09:17:08.2167910Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 30%] 2025-07-17T09:17:08.2168374Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 30%] 2025-07-17T09:17:08.2168945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 30%] 2025-07-17T09:17:08.2169454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 30%] 2025-07-17T09:17:08.2170106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2170688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2171272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2171859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2172442Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2173028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2173606Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2174193Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2174776Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2175350Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2176001Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2176674Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2177313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2177901Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2178479Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2179056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2179644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2180233Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2180810Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 30%] 2025-07-17T09:17:08.2181402Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2181989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2182570Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2183210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2183890Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2184582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2185158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2185735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2186313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2186887Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2187469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2188043Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2188616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2189202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2189781Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2190363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2191103Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2191736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2192323Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2192904Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2193488Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2194182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2194759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2195337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2195913Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2196500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2197085Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2197656Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2198419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2198968Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 31%] 2025-07-17T09:17:08.2199437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 31%] 2025-07-17T09:17:08.2199905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2200362Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 31%] 2025-07-17T09:17:08.2200831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 31%] 2025-07-17T09:17:08.2201289Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 31%] 2025-07-17T09:17:08.2201755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2202209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 31%] 2025-07-17T09:17:08.2202673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 31%] 2025-07-17T09:17:08.2203132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 31%] 2025-07-17T09:17:08.2203587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2204043Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 31%] 2025-07-17T09:17:08.2204501Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 31%] 2025-07-17T09:17:08.2205017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 31%] 2025-07-17T09:17:08.2205571Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2206076Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 31%] 2025-07-17T09:17:08.2206548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 31%] 2025-07-17T09:17:08.2207011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2207472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 31%] 2025-07-17T09:17:08.2207930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 31%] 2025-07-17T09:17:08.2208397Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2208859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 31%] 2025-07-17T09:17:08.2209320Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 31%] 2025-07-17T09:17:08.2209786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 31%] 2025-07-17T09:17:08.2210241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2210710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2211163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 31%] 2025-07-17T09:17:08.2211618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 31%] 2025-07-17T09:17:08.2212146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2212706Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 31%] 2025-07-17T09:17:08.2213213Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 31%] 2025-07-17T09:17:08.2213670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 31%] 2025-07-17T09:17:08.2214134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2214601Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 31%] 2025-07-17T09:17:08.2215063Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 31%] 2025-07-17T09:17:08.2215520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 31%] 2025-07-17T09:17:08.2215986Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2216448Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 31%] 2025-07-17T09:17:08.2216906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 31%] 2025-07-17T09:17:08.2217367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 31%] 2025-07-17T09:17:08.2217827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2218287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 31%] 2025-07-17T09:17:08.2218739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 31%] 2025-07-17T09:17:08.2219309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 31%] 2025-07-17T09:17:08.2219824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 31%] 2025-07-17T09:17:08.2220333Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 31%] 2025-07-17T09:17:08.2220793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 31%] 2025-07-17T09:17:08.2221250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 31%] 2025-07-17T09:17:08.2221843Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2222426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2223011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2223595Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2224180Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2224767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2225345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2225925Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2226616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2227260Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2227882Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2228458Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2229037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2229614Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2230183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2230759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2231342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2231930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2232517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2233094Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2233743Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2234503Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2235148Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2235735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2236316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2236897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2237470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2238045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2238626Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2239200Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2239785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2240363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2241006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2241702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2242329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2242909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2243504Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2244085Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2244667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 31%] 2025-07-17T09:17:08.2245245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2245822Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2246397Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2246974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2247551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2248184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2248875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2249497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2250069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2250538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 32%] 2025-07-17T09:17:08.2250996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 32%] 2025-07-17T09:17:08.2251458Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2251913Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2252379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0028s] [ 32%] 2025-07-17T09:17:08.2252840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 32%] 2025-07-17T09:17:08.2253299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2253759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2254214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0028s] [ 32%] 2025-07-17T09:17:08.2254671Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 32%] 2025-07-17T09:17:08.2255121Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2255701Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2256216Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0028s] [ 32%] 2025-07-17T09:17:08.2256720Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 32%] 2025-07-17T09:17:08.2257180Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2257634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2258104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 32%] 2025-07-17T09:17:08.2258564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 32%] 2025-07-17T09:17:08.2259030Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 32%] 2025-07-17T09:17:08.2259489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2259952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 32%] 2025-07-17T09:17:08.2260411Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2260872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2261335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2261786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2262240Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2262816Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2263321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2263828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2264283Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2264747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 32%] 2025-07-17T09:17:08.2265201Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2265662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2266125Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2266581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 32%] 2025-07-17T09:17:08.2267042Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2267501Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2267967Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2268425Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2268891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2269351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2269918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 32%] 2025-07-17T09:17:08.2270433Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2270933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2271392Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2271847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 32%] 2025-07-17T09:17:08.2272308Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2272765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 32%] 2025-07-17T09:17:08.2273352Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2274001Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2274589Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2275170Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2275749Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2276334Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2276984Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2277687Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2278331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2278906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2279478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2280046Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2280619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2281198Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2281771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2282340Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2282920Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2283500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2284180Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2284855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2285484Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2286072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2286650Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2287227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2287803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2288374Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2288943Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2289512Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2290096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2290673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2291305Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2291985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2292611Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2293197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2293782Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2294354Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2294935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2295515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2296097Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2296675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2297250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2297826Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2298394Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2299182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2299815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2300387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2300965Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2301534Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 32%] 2025-07-17T09:17:08.2302006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0048s] [ 32%] 2025-07-17T09:17:08.2302475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 32%] 2025-07-17T09:17:08.2302939Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0059s] [ 32%] 2025-07-17T09:17:08.2303405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0047s] [ 32%] 2025-07-17T09:17:08.2303872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 32%] 2025-07-17T09:17:08.2304346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 32%] 2025-07-17T09:17:08.2304812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0060s] [ 32%] 2025-07-17T09:17:08.2305277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0048s] [ 32%] 2025-07-17T09:17:08.2305786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 32%] 2025-07-17T09:17:08.2306352Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 32%] 2025-07-17T09:17:08.2306878Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0059s] [ 32%] 2025-07-17T09:17:08.2307335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0047s] [ 33%] 2025-07-17T09:17:08.2307803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 33%] 2025-07-17T09:17:08.2308263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 33%] 2025-07-17T09:17:08.2308730Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0059s] [ 33%] 2025-07-17T09:17:08.2309194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0047s] [ 33%] 2025-07-17T09:17:08.2309665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 33%] 2025-07-17T09:17:08.2310136Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2310600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0065s] [ 33%] 2025-07-17T09:17:08.2311065Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0051s] [ 33%] 2025-07-17T09:17:08.2311535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2312003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2312472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0065s] [ 33%] 2025-07-17T09:17:08.2313045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0051s] [ 33%] 2025-07-17T09:17:08.2313562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2314131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2314598Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0065s] [ 33%] 2025-07-17T09:17:08.2315058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0050s] [ 33%] 2025-07-17T09:17:08.2315527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2316005Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2316466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0065s] [ 33%] 2025-07-17T09:17:08.2316935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0050s] [ 33%] 2025-07-17T09:17:08.2317398Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2317866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2318328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0065s] [ 33%] 2025-07-17T09:17:08.2318794Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0051s] [ 33%] 2025-07-17T09:17:08.2319266Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2319732Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2320322Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0065s] [ 33%] 2025-07-17T09:17:08.2320847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0050s] [ 33%] 2025-07-17T09:17:08.2321366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 33%] 2025-07-17T09:17:08.2321825Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2322293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0065s] [ 33%] 2025-07-17T09:17:08.2322751Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0050s] [ 33%] 2025-07-17T09:17:08.2323217Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 33%] 2025-07-17T09:17:08.2323684Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2324150Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0065s] [ 33%] 2025-07-17T09:17:08.2324612Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0050s] [ 33%] 2025-07-17T09:17:08.2325201Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2325802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2326392Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2326971Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2327677Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2328318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2328954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2329548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2330132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2330715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0014s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2331291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2331869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2332460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2333044Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2333629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2334204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2334918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2335581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2336224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2336817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2337414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2338011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2338604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2339189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2339777Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2340359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2340946Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2341529Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2342227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2342873Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2343499Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2344085Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2344678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2345263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2345856Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2346437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2347030Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2347622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2348207Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2348804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2349527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2350160Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2350789Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2351372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2351959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2352548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2353128Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2353713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 33%] 2025-07-17T09:17:08.2354277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0054s] [ 33%] 2025-07-17T09:17:08.2354763Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 33%] 2025-07-17T09:17:08.2355227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0066s] [ 33%] 2025-07-17T09:17:08.2355696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0054s] [ 33%] 2025-07-17T09:17:08.2356176Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 33%] 2025-07-17T09:17:08.2356775Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 33%] 2025-07-17T09:17:08.2357313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0066s] [ 33%] 2025-07-17T09:17:08.2357837Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0053s] [ 33%] 2025-07-17T09:17:08.2358304Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0038s] [ 33%] 2025-07-17T09:17:08.2358767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 33%] 2025-07-17T09:17:08.2359234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0066s] [ 33%] 2025-07-17T09:17:08.2359697Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0053s] [ 33%] 2025-07-17T09:17:08.2360158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0038s] [ 33%] 2025-07-17T09:17:08.2360631Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 33%] 2025-07-17T09:17:08.2361090Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0066s] [ 33%] 2025-07-17T09:17:08.2361553Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0054s] [ 33%] 2025-07-17T09:17:08.2362025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 33%] 2025-07-17T09:17:08.2362498Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 33%] 2025-07-17T09:17:08.2362965Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0072s] [ 33%] 2025-07-17T09:17:08.2363425Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0058s] [ 33%] 2025-07-17T09:17:08.2364031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 33%] 2025-07-17T09:17:08.2364554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 33%] 2025-07-17T09:17:08.2365074Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0072s] [ 33%] 2025-07-17T09:17:08.2365541Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0058s] [ 33%] 2025-07-17T09:17:08.2366012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0041s] [ 33%] 2025-07-17T09:17:08.2366485Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0041s] [ 33%] 2025-07-17T09:17:08.2366947Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0072s] [ 33%] 2025-07-17T09:17:08.2367413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0058s] [ 33%] 2025-07-17T09:17:08.2367877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0041s] [ 33%] 2025-07-17T09:17:08.2368348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0041s] [ 33%] 2025-07-17T09:17:08.2368809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0072s] [ 34%] 2025-07-17T09:17:08.2369274Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0058s] [ 34%] 2025-07-17T09:17:08.2369744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 34%] 2025-07-17T09:17:08.2370214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 34%] 2025-07-17T09:17:08.2370679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0074s] [ 34%] 2025-07-17T09:17:08.2371293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0057s] [ 34%] 2025-07-17T09:17:08.2371821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 34%] 2025-07-17T09:17:08.2372284Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 34%] 2025-07-17T09:17:08.2372760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0073s] [ 34%] 2025-07-17T09:17:08.2373232Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0058s] [ 34%] 2025-07-17T09:17:08.2373688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0041s] [ 34%] 2025-07-17T09:17:08.2374155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0040s] [ 34%] 2025-07-17T09:17:08.2374615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0072s] [ 34%] 2025-07-17T09:17:08.2375078Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0057s] [ 34%] 2025-07-17T09:17:08.2375543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0041s] [ 34%] 2025-07-17T09:17:08.2376011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0040s] [ 34%] 2025-07-17T09:17:08.2376477Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0072s] [ 34%] 2025-07-17T09:17:08.2376936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0057s] [ 34%] 2025-07-17T09:17:08.2377528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2378217Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2378862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2379505Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2380095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2380683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2381266Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2381854Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2382435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2383013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2383595Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2384171Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2384757Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2385455Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2386085Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2386713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2387305Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2387894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2388479Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2389065Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2389659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2390244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2390840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2391428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2392009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2392707Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2393338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2394026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2394625Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2395205Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2395788Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2396369Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2396959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2397548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2398131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2398722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2399310Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2400028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2400684Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2401324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2401914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2402500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2403091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2403678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2404259Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2404843Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2405429Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2406004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2406475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0047s] [ 34%] 2025-07-17T09:17:08.2406989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 34%] 2025-07-17T09:17:08.2407556Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 34%] 2025-07-17T09:17:08.2408064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 34%] 2025-07-17T09:17:08.2408531Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 34%] 2025-07-17T09:17:08.2409000Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 34%] 2025-07-17T09:17:08.2409460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 34%] 2025-07-17T09:17:08.2409927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 34%] 2025-07-17T09:17:08.2410387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 34%] 2025-07-17T09:17:08.2410849Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 34%] 2025-07-17T09:17:08.2411300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 34%] 2025-07-17T09:17:08.2411758Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 34%] 2025-07-17T09:17:08.2412222Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 34%] 2025-07-17T09:17:08.2412687Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 34%] 2025-07-17T09:17:08.2413147Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 34%] 2025-07-17T09:17:08.2413601Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 34%] 2025-07-17T09:17:08.2414122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 34%] 2025-07-17T09:17:08.2414742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 34%] 2025-07-17T09:17:08.2415275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 34%] 2025-07-17T09:17:08.2415739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 34%] 2025-07-17T09:17:08.2416207Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 34%] 2025-07-17T09:17:08.2416677Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 34%] 2025-07-17T09:17:08.2417140Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 34%] 2025-07-17T09:17:08.2417606Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 34%] 2025-07-17T09:17:08.2418067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 34%] 2025-07-17T09:17:08.2418531Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 34%] 2025-07-17T09:17:08.2418989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 34%] 2025-07-17T09:17:08.2419439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 34%] 2025-07-17T09:17:08.2419904Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 34%] 2025-07-17T09:17:08.2420362Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 34%] 2025-07-17T09:17:08.2420819Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 34%] 2025-07-17T09:17:08.2421332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 34%] 2025-07-17T09:17:08.2421909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 34%] 2025-07-17T09:17:08.2422424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 34%] 2025-07-17T09:17:08.2422885Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 34%] 2025-07-17T09:17:08.2423357Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 34%] 2025-07-17T09:17:08.2423827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 34%] 2025-07-17T09:17:08.2424297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 34%] 2025-07-17T09:17:08.2424759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 34%] 2025-07-17T09:17:08.2425230Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 34%] 2025-07-17T09:17:08.2425694Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 34%] 2025-07-17T09:17:08.2426152Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 34%] 2025-07-17T09:17:08.2426617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 34%] 2025-07-17T09:17:08.2427073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 34%] 2025-07-17T09:17:08.2427537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 34%] 2025-07-17T09:17:08.2427992Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 34%] 2025-07-17T09:17:08.2428565Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 34%] 2025-07-17T09:17:08.2429080Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 34%] 2025-07-17T09:17:08.2429712Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2430303Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 34%] 2025-07-17T09:17:08.2430888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2431471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2432068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2432658Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2433241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2433819Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2434472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2435056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2435629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2436390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2437028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2437613Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2438195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2438770Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2439358Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2439940Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2440526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2441105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2441693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2442290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2442872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2443634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2444265Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2444842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2445426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2446000Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2446584Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2447167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2447745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2448325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2448910Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2449494Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2450076Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2450808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2451442Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2452025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2452616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2453207Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2453779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2454361Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2454936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2455515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2456100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2456675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2457254Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2457984Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_143_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2458514Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0055s] [ 35%] 2025-07-17T09:17:08.2458985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0049s] [ 35%] 2025-07-17T09:17:08.2459458Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0116s] [ 35%] 2025-07-17T09:17:08.2459929Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0097s] [ 35%] 2025-07-17T09:17:08.2460404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0050s] [ 35%] 2025-07-17T09:17:08.2460877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0048s] [ 35%] 2025-07-17T09:17:08.2461346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0117s] [ 35%] 2025-07-17T09:17:08.2461819Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0097s] [ 35%] 2025-07-17T09:17:08.2462286Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0049s] [ 35%] 2025-07-17T09:17:08.2462745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0048s] [ 35%] 2025-07-17T09:17:08.2463212Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0116s] [ 35%] 2025-07-17T09:17:08.2463675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0097s] [ 35%] 2025-07-17T09:17:08.2464144Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0049s] [ 35%] 2025-07-17T09:17:08.2464660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0049s] [ 35%] 2025-07-17T09:17:08.2465237Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0116s] [ 35%] 2025-07-17T09:17:08.2465754Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0096s] [ 35%] 2025-07-17T09:17:08.2466225Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0053s] [ 35%] 2025-07-17T09:17:08.2466700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0053s] [ 35%] 2025-07-17T09:17:08.2467171Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0128s] [ 35%] 2025-07-17T09:17:08.2467647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0108s] [ 35%] 2025-07-17T09:17:08.2468117Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0052s] [ 35%] 2025-07-17T09:17:08.2468597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0052s] [ 35%] 2025-07-17T09:17:08.2469073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0128s] [ 35%] 2025-07-17T09:17:08.2469538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0106s] [ 35%] 2025-07-17T09:17:08.2470009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0053s] [ 35%] 2025-07-17T09:17:08.2470476Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0052s] [ 35%] 2025-07-17T09:17:08.2470942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0128s] [ 35%] 2025-07-17T09:17:08.2471399Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0106s] [ 35%] 2025-07-17T09:17:08.2471990Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0053s] [ 35%] 2025-07-17T09:17:08.2472519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0052s] [ 35%] 2025-07-17T09:17:08.2473034Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0128s] [ 35%] 2025-07-17T09:17:08.2473501Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0106s] [ 35%] 2025-07-17T09:17:08.2474068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0053s] [ 35%] 2025-07-17T09:17:08.2474545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0052s] [ 35%] 2025-07-17T09:17:08.2475013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0129s] [ 35%] 2025-07-17T09:17:08.2475488Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0107s] [ 35%] 2025-07-17T09:17:08.2475965Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0053s] [ 35%] 2025-07-17T09:17:08.2476435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0052s] [ 35%] 2025-07-17T09:17:08.2476907Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0128s] [ 35%] 2025-07-17T09:17:08.2477372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0108s] [ 35%] 2025-07-17T09:17:08.2477847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0053s] [ 35%] 2025-07-17T09:17:08.2478312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0052s] [ 35%] 2025-07-17T09:17:08.2478779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0127s] [ 35%] 2025-07-17T09:17:08.2479379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0105s] [ 35%] 2025-07-17T09:17:08.2479907Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0052s] [ 35%] 2025-07-17T09:17:08.2480434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0052s] [ 35%] 2025-07-17T09:17:08.2480895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0128s] [ 35%] 2025-07-17T09:17:08.2481363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0107s] [ 35%] 2025-07-17T09:17:08.2481961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2482555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2483147Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2483736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2484328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2484921Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2485507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2486095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2486882Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2487520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2488097Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2488676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2489262Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2489843Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2490434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2491017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2491607Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2492199Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2492785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2493373Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2494123Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2494769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 35%] 2025-07-17T09:17:08.2495363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2495956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2496547Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2497132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2497715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2498310Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2498894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2499486Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2500071Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2500650Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2501418Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2502070Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2502660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2503257Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2503847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2504438Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2505026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2505615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2506206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2506790Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2507383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2507962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2508704Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2509337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2509919Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2510512Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2510983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0053s] [ 36%] 2025-07-17T09:17:08.2511455Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0048s] [ 36%] 2025-07-17T09:17:08.2511929Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0124s] [ 36%] 2025-07-17T09:17:08.2512393Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0089s] [ 36%] 2025-07-17T09:17:08.2512868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0048s] [ 36%] 2025-07-17T09:17:08.2513335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0047s] [ 36%] 2025-07-17T09:17:08.2513809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0123s] [ 36%] 2025-07-17T09:17:08.2514348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0092s] [ 36%] 2025-07-17T09:17:08.2514829Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0048s] [ 36%] 2025-07-17T09:17:08.2515291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0049s] [ 36%] 2025-07-17T09:17:08.2515945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0126s] [ 36%] 2025-07-17T09:17:08.2516472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0091s] [ 36%] 2025-07-17T09:17:08.2516936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0050s] [ 36%] 2025-07-17T09:17:08.2517409Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0050s] [ 36%] 2025-07-17T09:17:08.2517875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0125s] [ 36%] 2025-07-17T09:17:08.2518348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0090s] [ 36%] 2025-07-17T09:17:08.2518823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0052s] [ 36%] 2025-07-17T09:17:08.2519293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2519762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0134s] [ 36%] 2025-07-17T09:17:08.2520227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0098s] [ 36%] 2025-07-17T09:17:08.2520702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0052s] [ 36%] 2025-07-17T09:17:08.2521179Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2521649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0134s] [ 36%] 2025-07-17T09:17:08.2522119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0097s] [ 36%] 2025-07-17T09:17:08.2522633Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2523202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0052s] [ 36%] 2025-07-17T09:17:08.2523728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0134s] [ 36%] 2025-07-17T09:17:08.2524202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0097s] [ 36%] 2025-07-17T09:17:08.2524673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2525146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2525610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0134s] [ 36%] 2025-07-17T09:17:08.2526072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0097s] [ 36%] 2025-07-17T09:17:08.2526545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2527012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2527488Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0134s] [ 36%] 2025-07-17T09:17:08.2527952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0097s] [ 36%] 2025-07-17T09:17:08.2528429Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2528915Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2529383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0134s] [ 36%] 2025-07-17T09:17:08.2530031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0097s] [ 36%] 2025-07-17T09:17:08.2530550Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2531069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2531534Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0134s] [ 36%] 2025-07-17T09:17:08.2532003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0097s] [ 36%] 2025-07-17T09:17:08.2532471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2532936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0051s] [ 36%] 2025-07-17T09:17:08.2533400Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0134s] [ 36%] 2025-07-17T09:17:08.2533868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0097s] [ 36%] 2025-07-17T09:17:08.2534465Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2535055Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2535644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2536231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2536821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2537566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2538207Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2538802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2539390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2539970Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2540558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2541144Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2541728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2542314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2542897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2543483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2544081Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2544819Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2545459Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2546044Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2546645Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2547239Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2547822Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2548418Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2549001Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2549588Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2550172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2550752Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2551340Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2552096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2552739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2553324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2553989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2554600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2555188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2555798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2556399Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2556988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2557587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2558175Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2558761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2559546Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 36%] 2025-07-17T09:17:08.2560188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2560773Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2561364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2561954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2562540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2563120Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2563600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 37%] 2025-07-17T09:17:08.2564064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 37%] 2025-07-17T09:17:08.2564533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0054s] [ 37%] 2025-07-17T09:17:08.2564999Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0048s] [ 37%] 2025-07-17T09:17:08.2565471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 37%] 2025-07-17T09:17:08.2565941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 37%] 2025-07-17T09:17:08.2566511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0054s] [ 37%] 2025-07-17T09:17:08.2567034Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0047s] [ 37%] 2025-07-17T09:17:08.2567545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 37%] 2025-07-17T09:17:08.2568009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 37%] 2025-07-17T09:17:08.2568480Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0053s] [ 37%] 2025-07-17T09:17:08.2568941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0047s] [ 37%] 2025-07-17T09:17:08.2569403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 37%] 2025-07-17T09:17:08.2569865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 37%] 2025-07-17T09:17:08.2570334Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0053s] [ 37%] 2025-07-17T09:17:08.2570791Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0046s] [ 37%] 2025-07-17T09:17:08.2571263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 37%] 2025-07-17T09:17:08.2571733Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 37%] 2025-07-17T09:17:08.2572198Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0062s] [ 37%] 2025-07-17T09:17:08.2572665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0052s] [ 37%] 2025-07-17T09:17:08.2573131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2573715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2574233Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0062s] [ 37%] 2025-07-17T09:17:08.2574750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0053s] [ 37%] 2025-07-17T09:17:08.2575223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2575688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2576152Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0062s] [ 37%] 2025-07-17T09:17:08.2576606Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0052s] [ 37%] 2025-07-17T09:17:08.2577074Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2577537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2578006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0062s] [ 37%] 2025-07-17T09:17:08.2578470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0053s] [ 37%] 2025-07-17T09:17:08.2578935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 37%] 2025-07-17T09:17:08.2579404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2579866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0064s] [ 37%] 2025-07-17T09:17:08.2580333Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0053s] [ 37%] 2025-07-17T09:17:08.2580983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2581508Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2581985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0063s] [ 37%] 2025-07-17T09:17:08.2582452Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0053s] [ 37%] 2025-07-17T09:17:08.2582922Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2583383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2583847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0062s] [ 37%] 2025-07-17T09:17:08.2584306Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0052s] [ 37%] 2025-07-17T09:17:08.2584769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2585236Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 37%] 2025-07-17T09:17:08.2585702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0062s] [ 37%] 2025-07-17T09:17:08.2586168Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0052s] [ 37%] 2025-07-17T09:17:08.2586760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2587351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2588057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2588695Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2589337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2589924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2590516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2591101Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2591685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2592266Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2592843Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2593428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2594077Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2594654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2595368Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2596011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2596661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2597253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2597835Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2598422Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2599010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2599602Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2600189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2600772Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2601359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2601940Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2602629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2603262Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2603891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2604479Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2605073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2605660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2606251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2606834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2607424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2608003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2608599Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2609194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2609896Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2610533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2611162Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2611752Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2612331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2612909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2613493Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2614073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2614656Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2615236Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 37%] 2025-07-17T09:17:08.2615713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.1889s] [ 37%] 2025-07-17T09:17:08.2616197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0242s] [ 37%] 2025-07-17T09:17:08.2616668Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0764s] [ 37%] 2025-07-17T09:17:08.2617248Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0515s] [ 37%] 2025-07-17T09:17:08.2617774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0246s] [ 37%] 2025-07-17T09:17:08.2618296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0238s] [ 37%] 2025-07-17T09:17:08.2618776Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0765s] [ 37%] 2025-07-17T09:17:08.2619245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0524s] [ 37%] 2025-07-17T09:17:08.2619720Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0257s] [ 37%] 2025-07-17T09:17:08.2620189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0258s] [ 37%] 2025-07-17T09:17:08.2620659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0785s] [ 37%] 2025-07-17T09:17:08.2621127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0516s] [ 37%] 2025-07-17T09:17:08.2621592Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.1547s] [ 37%] 2025-07-17T09:17:08.2622060Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1041s] [ 37%] 2025-07-17T09:17:08.2622525Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.1383s] [ 38%] 2025-07-17T09:17:08.2622991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3208s] [ 38%] 2025-07-17T09:17:08.2623462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.1756s] [ 38%] 2025-07-17T09:17:08.2623990Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1925s] [ 38%] 2025-07-17T09:17:08.2624565Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.2060s] [ 38%] 2025-07-17T09:17:08.2625085Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.3952s] [ 38%] 2025-07-17T09:17:08.2625567Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.2483s] [ 38%] 2025-07-17T09:17:08.2626041Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0307s] [ 38%] 2025-07-17T09:17:08.2626518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.6895s] [ 38%] 2025-07-17T09:17:08.2626991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.3955s] [ 38%] 2025-07-17T09:17:08.2627462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.3798s] [ 38%] 2025-07-17T09:17:08.2627937Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.2244s] [ 38%] 2025-07-17T09:17:08.2628401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.4922s] [ 38%] 2025-07-17T09:17:08.2628869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.5866s] [ 38%] 2025-07-17T09:17:08.2629338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.4163s] [ 38%] 2025-07-17T09:17:08.2629814Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.2538s] [ 38%] 2025-07-17T09:17:08.2630278Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.3805s] [ 38%] 2025-07-17T09:17:08.2630748Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.6846s] [ 38%] 2025-07-17T09:17:08.2631336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.4498s] [ 38%] 2025-07-17T09:17:08.2631859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1955s] [ 38%] 2025-07-17T09:17:08.2632399Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.3075s] [ 38%] 2025-07-17T09:17:08.2632865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.7495s] [ 38%] 2025-07-17T09:17:08.2633346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.4115s] [ 38%] 2025-07-17T09:17:08.2633817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1954s] [ 38%] 2025-07-17T09:17:08.2634363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.2845s] [ 38%] 2025-07-17T09:17:08.2634840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.7494s] [ 38%] 2025-07-17T09:17:08.2635315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.4082s] [ 38%] 2025-07-17T09:17:08.2635787Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1960s] [ 38%] 2025-07-17T09:17:08.2636271Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.2856s] [ 38%] 2025-07-17T09:17:08.2636745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.7488s] [ 38%] 2025-07-17T09:17:08.2637230Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.4103s] [ 38%] 2025-07-17T09:17:08.2637712Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1956s] [ 38%] 2025-07-17T09:17:08.2638187Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.2842s] [ 38%] 2025-07-17T09:17:08.2638779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.7390s] [ 38%] 2025-07-17T09:17:08.2639324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.4143s] [ 38%] 2025-07-17T09:17:08.2639913Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1743s] [ 38%] 2025-07-17T09:17:08.2640386Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.1662s] [ 38%] 2025-07-17T09:17:08.2640855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4517s] [ 38%] 2025-07-17T09:17:08.2641329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3207s] [ 38%] 2025-07-17T09:17:08.2641802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1575s] [ 38%] 2025-07-17T09:17:08.2642268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.2605s] [ 38%] 2025-07-17T09:17:08.2642744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4322s] [ 38%] 2025-07-17T09:17:08.2643208Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.2480s] [ 38%] 2025-07-17T09:17:08.2643673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1547s] [ 38%] 2025-07-17T09:17:08.2644133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.3448s] [ 38%] 2025-07-17T09:17:08.2644597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3877s] [ 38%] 2025-07-17T09:17:08.2645068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.2923s] [ 38%] 2025-07-17T09:17:08.2645528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1086s] [ 38%] 2025-07-17T09:17:08.2646109Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.4048s] [ 38%] 2025-07-17T09:17:08.2646620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.5443s] [ 38%] 2025-07-17T09:17:08.2647141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3171s] [ 38%] 2025-07-17T09:17:08.2647609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.2054s] [ 38%] 2025-07-17T09:17:08.2648082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.4535s] [ 38%] 2025-07-17T09:17:08.2648551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.5885s] [ 38%] 2025-07-17T09:17:08.2649023Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.4424s] [ 38%] 2025-07-17T09:17:08.2649498Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.2279s] [ 38%] 2025-07-17T09:17:08.2649971Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.4121s] [ 38%] 2025-07-17T09:17:08.2650450Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.6653s] [ 38%] 2025-07-17T09:17:08.2650914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.3281s] [ 38%] 2025-07-17T09:17:08.2651388Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.2605s] [ 38%] 2025-07-17T09:17:08.2651851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.3583s] [ 38%] 2025-07-17T09:17:08.2652316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.6648s] [ 38%] 2025-07-17T09:17:08.2652840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.4445s] [ 38%] 2025-07-17T09:17:08.2653404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.2450s] [ 38%] 2025-07-17T09:17:08.2653926Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.3019s] [ 38%] 2025-07-17T09:17:08.2654388Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.5568s] [ 38%] 2025-07-17T09:17:08.2654870Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.4888s] [ 38%] 2025-07-17T09:17:08.2655343Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1966s] [ 38%] 2025-07-17T09:17:08.2655810Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.3860s] [ 38%] 2025-07-17T09:17:08.2656283Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.6792s] [ 38%] 2025-07-17T09:17:08.2656760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.4424s] [ 38%] 2025-07-17T09:17:08.2657235Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1974s] [ 38%] 2025-07-17T09:17:08.2657709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.3244s] [ 38%] 2025-07-17T09:17:08.2658185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.7122s] [ 38%] 2025-07-17T09:17:08.2658662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.4247s] [ 38%] 2025-07-17T09:17:08.2659126Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1970s] [ 38%] 2025-07-17T09:17:08.2659595Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.3092s] [ 38%] 2025-07-17T09:17:08.2660157Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.7176s] [ 38%] 2025-07-17T09:17:08.2660705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.4197s] [ 38%] 2025-07-17T09:17:08.2661218Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1962s] [ 38%] 2025-07-17T09:17:08.2661697Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.3091s] [ 38%] 2025-07-17T09:17:08.2662169Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.7391s] [ 38%] 2025-07-17T09:17:08.2662641Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3907s] [ 38%] 2025-07-17T09:17:08.2663122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1244s] [ 38%] 2025-07-17T09:17:08.2663585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.2329s] [ 38%] 2025-07-17T09:17:08.2664061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.5192s] [ 38%] 2025-07-17T09:17:08.2664540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3087s] [ 38%] 2025-07-17T09:17:08.2665017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1692s] [ 38%] 2025-07-17T09:17:08.2665495Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.3561s] [ 38%] 2025-07-17T09:17:08.2665962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4397s] [ 38%] 2025-07-17T09:17:08.2666432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.2837s] [ 38%] 2025-07-17T09:17:08.2666895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1396s] [ 38%] 2025-07-17T09:17:08.2667504Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.5108s] [ 38%] 2025-07-17T09:17:08.2668021Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.4244s] [ 38%] 2025-07-17T09:17:08.2668544Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.1990s] [ 38%] 2025-07-17T09:17:08.2669013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1001s] [ 38%] 2025-07-17T09:17:08.2669486Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.6543s] [ 38%] 2025-07-17T09:17:08.2669956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3520s] [ 38%] 2025-07-17T09:17:08.2670423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.1541s] [ 38%] 2025-07-17T09:17:08.2670900Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.2940s] [ 38%] 2025-07-17T09:17:08.2671370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.7580s] [ 38%] 2025-07-17T09:17:08.2671844Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4902s] [ 38%] 2025-07-17T09:17:08.2672324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3921s] [ 38%] 2025-07-17T09:17:08.2672797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.2090s] [ 38%] 2025-07-17T09:17:08.2673273Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.5713s] [ 38%] 2025-07-17T09:17:08.2673743Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4692s] [ 38%] 2025-07-17T09:17:08.2674324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.4262s] [ 38%] 2025-07-17T09:17:08.2674931Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0800s] [ 38%] 2025-07-17T09:17:08.2675463Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.7054s] [ 38%] 2025-07-17T09:17:08.2675989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.5269s] [ 38%] 2025-07-17T09:17:08.2676459Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.2831s] [ 38%] 2025-07-17T09:17:08.2676936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.2079s] [ 38%] 2025-07-17T09:17:08.2677401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.6701s] [ 38%] 2025-07-17T09:17:08.2677869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.5472s] [ 38%] 2025-07-17T09:17:08.2678341Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.4308s] [ 38%] 2025-07-17T09:17:08.2678817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0918s] [ 38%] 2025-07-17T09:17:08.2679290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.6902s] [ 39%] 2025-07-17T09:17:08.2679761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4223s] [ 39%] 2025-07-17T09:17:08.2680239Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3996s] [ 39%] 2025-07-17T09:17:08.2680710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1997s] [ 39%] 2025-07-17T09:17:08.2681185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.6653s] [ 39%] 2025-07-17T09:17:08.2681716Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.3971s] [ 39%] 2025-07-17T09:17:08.2682287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.4494s] [ 39%] 2025-07-17T09:17:08.2682810Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1947s] [ 39%] 2025-07-17T09:17:08.2683273Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.6457s] [ 39%] 2025-07-17T09:17:08.2683750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3970s] [ 39%] 2025-07-17T09:17:08.2684218Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.4841s] [ 39%] 2025-07-17T09:17:08.2684690Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1995s] [ 39%] 2025-07-17T09:17:08.2685159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.6055s] [ 39%] 2025-07-17T09:17:08.2685626Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3964s] [ 39%] 2025-07-17T09:17:08.2686098Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.4709s] [ 39%] 2025-07-17T09:17:08.2686563Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1497s] [ 39%] 2025-07-17T09:17:08.2687030Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.3995s] [ 39%] 2025-07-17T09:17:08.2687498Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4193s] [ 39%] 2025-07-17T09:17:08.2687972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.2792s] [ 39%] 2025-07-17T09:17:08.2688443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1566s] [ 39%] 2025-07-17T09:17:08.2689032Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.5107s] [ 39%] 2025-07-17T09:17:08.2689561Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4290s] [ 39%] 2025-07-17T09:17:08.2690069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.3322s] [ 39%] 2025-07-17T09:17:08.2690536Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0817s] [ 39%] 2025-07-17T09:17:08.2691006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.5536s] [ 39%] 2025-07-17T09:17:08.2691473Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3455s] [ 39%] 2025-07-17T09:17:08.2691944Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.3221s] [ 39%] 2025-07-17T09:17:08.2692406Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1090s] [ 39%] 2025-07-17T09:17:08.2692875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.5767s] [ 39%] 2025-07-17T09:17:08.2693334Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3880s] [ 39%] 2025-07-17T09:17:08.2693816Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.2933s] [ 39%] 2025-07-17T09:17:08.2694288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.2335s] [ 39%] 2025-07-17T09:17:08.2694758Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.7188s] [ 39%] 2025-07-17T09:17:08.2695229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4234s] [ 39%] 2025-07-17T09:17:08.2695698Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3372s] [ 39%] 2025-07-17T09:17:08.2696276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.2044s] [ 39%] 2025-07-17T09:17:08.2696804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.6915s] [ 39%] 2025-07-17T09:17:08.2697336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4371s] [ 39%] 2025-07-17T09:17:08.2706089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.3964s] [ 39%] 2025-07-17T09:17:08.2706634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1731s] [ 39%] 2025-07-17T09:17:08.2707128Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.6656s] [ 39%] 2025-07-17T09:17:08.2707593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.4909s] [ 39%] 2025-07-17T09:17:08.2708069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.3840s] [ 39%] 2025-07-17T09:17:08.2708544Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1558s] [ 39%] 2025-07-17T09:17:08.2709010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.6425s] [ 39%] 2025-07-17T09:17:08.2709475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.5009s] [ 39%] 2025-07-17T09:17:08.2709947Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3694s] [ 39%] 2025-07-17T09:17:08.2710423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.2013s] [ 39%] 2025-07-17T09:17:08.2710895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.6769s] [ 39%] 2025-07-17T09:17:08.2711364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4288s] [ 39%] 2025-07-17T09:17:08.2712015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.4756s] [ 39%] 2025-07-17T09:17:08.2712545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.2012s] [ 39%] 2025-07-17T09:17:08.2713075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.5854s] [ 39%] 2025-07-17T09:17:08.2713545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4185s] [ 39%] 2025-07-17T09:17:08.2714091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.4460s] [ 39%] 2025-07-17T09:17:08.2714558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.2009s] [ 39%] 2025-07-17T09:17:08.2715025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.6257s] [ 39%] 2025-07-17T09:17:08.2715492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.4181s] [ 39%] 2025-07-17T09:17:08.2715958Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.4848s] [ 39%] 2025-07-17T09:17:08.2716427Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.2115s] [ 39%] 2025-07-17T09:17:08.2716889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.5766s] [ 39%] 2025-07-17T09:17:08.2717358Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.4304s] [ 39%] 2025-07-17T09:17:08.2717825Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.4280s] [ 39%] 2025-07-17T09:17:08.2718297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0988s] [ 39%] 2025-07-17T09:17:08.2718842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.1690s] [ 39%] 2025-07-17T09:17:08.2719428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.3743s] [ 39%] 2025-07-17T09:17:08.2719957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.2224s] [ 39%] 2025-07-17T09:17:08.2720420Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1063s] [ 39%] 2025-07-17T09:17:08.2720900Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.2652s] [ 39%] 2025-07-17T09:17:08.2721368Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.2813s] [ 39%] 2025-07-17T09:17:08.2721838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.2617s] [ 39%] 2025-07-17T09:17:08.2722301Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1317s] [ 39%] 2025-07-17T09:17:08.2722766Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.2697s] [ 39%] 2025-07-17T09:17:08.2723235Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3324s] [ 39%] 2025-07-17T09:17:08.2723696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.2335s] [ 39%] 2025-07-17T09:17:08.2724158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1271s] [ 39%] 2025-07-17T09:17:08.2724619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.3014s] [ 39%] 2025-07-17T09:17:08.2725088Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3706s] [ 39%] 2025-07-17T09:17:08.2725558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.2551s] [ 39%] 2025-07-17T09:17:08.2726102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1644s] [ 39%] 2025-07-17T09:17:08.2726673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.5066s] [ 39%] 2025-07-17T09:17:08.2727185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.3152s] [ 39%] 2025-07-17T09:17:08.2727660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3768s] [ 39%] 2025-07-17T09:17:08.2728129Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1586s] [ 39%] 2025-07-17T09:17:08.2728599Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.4313s] [ 39%] 2025-07-17T09:17:08.2729072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4051s] [ 39%] 2025-07-17T09:17:08.2729535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.3790s] [ 39%] 2025-07-17T09:17:08.2730004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1400s] [ 39%] 2025-07-17T09:17:08.2730466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.4912s] [ 39%] 2025-07-17T09:17:08.2730930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3348s] [ 39%] 2025-07-17T09:17:08.2731395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.4147s] [ 39%] 2025-07-17T09:17:08.2731866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1355s] [ 39%] 2025-07-17T09:17:08.2732335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.4191s] [ 39%] 2025-07-17T09:17:08.2732793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.4070s] [ 39%] 2025-07-17T09:17:08.2733381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.2987s] [ 39%] 2025-07-17T09:17:08.2733908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1247s] [ 39%] 2025-07-17T09:17:08.2734430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.5242s] [ 39%] 2025-07-17T09:17:08.2734892Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4688s] [ 39%] 2025-07-17T09:17:08.2735372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.1952s] [ 39%] 2025-07-17T09:17:08.2735842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1691s] [ 39%] 2025-07-17T09:17:08.2736308Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.5553s] [ 39%] 2025-07-17T09:17:08.2736777Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4024s] [ 39%] 2025-07-17T09:17:08.2737241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.2321s] [ 39%] 2025-07-17T09:17:08.2737711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0595s] [ 39%] 2025-07-17T09:17:08.2738177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.6657s] [ 39%] 2025-07-17T09:17:08.2738648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.4024s] [ 39%] 2025-07-17T09:17:08.2739122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.2318s] [ 39%] 2025-07-17T09:17:08.2739585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1686s] [ 39%] 2025-07-17T09:17:08.2740050Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.5551s] [ 39%] 2025-07-17T09:17:08.2740624Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.4023s] [ 39%] 2025-07-17T09:17:08.2741147Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.2444s] [ 39%] 2025-07-17T09:17:08.2741657Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1319s] [ 39%] 2025-07-17T09:17:08.2742123Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.2834s] [ 39%] 2025-07-17T09:17:08.2742600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.3538s] [ 39%] 2025-07-17T09:17:08.2743067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3205s] [ 39%] 2025-07-17T09:17:08.2743533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1073s] [ 39%] 2025-07-17T09:17:08.2743994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.2125s] [ 40%] 2025-07-17T09:17:08.2744463Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.3413s] [ 40%] 2025-07-17T09:17:08.2744922Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.2457s] [ 40%] 2025-07-17T09:17:08.2745383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0808s] [ 40%] 2025-07-17T09:17:08.2745849Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.3107s] [ 40%] 2025-07-17T09:17:08.2746303Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3255s] [ 40%] 2025-07-17T09:17:08.2746768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.2796s] [ 40%] 2025-07-17T09:17:08.2747223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1180s] [ 40%] 2025-07-17T09:17:08.2747798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.1842s] [ 40%] 2025-07-17T09:17:08.2748316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3721s] [ 40%] 2025-07-17T09:17:08.2748856Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.2252s] [ 40%] 2025-07-17T09:17:08.2749324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1597s] [ 40%] 2025-07-17T09:17:08.2749789Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.4971s] [ 40%] 2025-07-17T09:17:08.2750257Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.3903s] [ 40%] 2025-07-17T09:17:08.2750725Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3117s] [ 40%] 2025-07-17T09:17:08.2751197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1598s] [ 40%] 2025-07-17T09:17:08.2751664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.4956s] [ 40%] 2025-07-17T09:17:08.2752129Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.3903s] [ 40%] 2025-07-17T09:17:08.2752590Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.2739s] [ 40%] 2025-07-17T09:17:08.2753052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1801s] [ 40%] 2025-07-17T09:17:08.2753516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.5487s] [ 40%] 2025-07-17T09:17:08.2754042Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.4008s] [ 40%] 2025-07-17T09:17:08.2754511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.2689s] [ 40%] 2025-07-17T09:17:08.2755102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1644s] [ 40%] 2025-07-17T09:17:08.2755633Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.4853s] [ 40%] 2025-07-17T09:17:08.2756154Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.3960s] [ 40%] 2025-07-17T09:17:08.2756618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3376s] [ 40%] 2025-07-17T09:17:08.2757091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0818s] [ 40%] 2025-07-17T09:17:08.2757556Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.5512s] [ 40%] 2025-07-17T09:17:08.2758018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4492s] [ 40%] 2025-07-17T09:17:08.2758487Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.3349s] [ 40%] 2025-07-17T09:17:08.2758961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.1695s] [ 40%] 2025-07-17T09:17:08.2759435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.4175s] [ 40%] 2025-07-17T09:17:08.2759897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.4057s] [ 40%] 2025-07-17T09:17:08.2760367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.3281s] [ 40%] 2025-07-17T09:17:08.2760830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1696s] [ 40%] 2025-07-17T09:17:08.2761295Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.4576s] [ 40%] 2025-07-17T09:17:08.2761747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.4053s] [ 40%] 2025-07-17T09:17:08.2762436Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.3766s] [ 40%] 2025-07-17T09:17:08.2762957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.1556s] [ 40%] 2025-07-17T09:17:08.2763417Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.4259s] [ 40%] 2025-07-17T09:17:08.2763879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.4060s] [ 40%] 2025-07-17T09:17:08.2764343Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0061s] [ 40%] 2025-07-17T09:17:08.2764818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0049s] [ 40%] 2025-07-17T09:17:08.2765277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0715s] [ 40%] 2025-07-17T09:17:08.2765745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0505s] [ 40%] 2025-07-17T09:17:08.2766214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0479s] [ 40%] 2025-07-17T09:17:08.2766676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0052s] [ 40%] 2025-07-17T09:17:08.2767141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0961s] [ 40%] 2025-07-17T09:17:08.2767604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0488s] [ 40%] 2025-07-17T09:17:08.2768070Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0054s] [ 40%] 2025-07-17T09:17:08.2768536Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0050s] [ 40%] 2025-07-17T09:17:08.2769054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0760s] [ 40%] 2025-07-17T09:17:08.2769616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0592s] [ 40%] 2025-07-17T09:17:08.2770124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0055s] [ 40%] 2025-07-17T09:17:08.2770587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0126s] [ 40%] 2025-07-17T09:17:08.2771052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0794s] [ 40%] 2025-07-17T09:17:08.2771519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0600s] [ 40%] 2025-07-17T09:17:08.2771984Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0176s] [ 40%] 2025-07-17T09:17:08.2772454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0193s] [ 40%] 2025-07-17T09:17:08.2772923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0702s] [ 40%] 2025-07-17T09:17:08.2773387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0600s] [ 40%] 2025-07-17T09:17:08.2773864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0377s] [ 40%] 2025-07-17T09:17:08.2774330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0053s] [ 40%] 2025-07-17T09:17:08.2774801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0788s] [ 40%] 2025-07-17T09:17:08.2775264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0795s] [ 40%] 2025-07-17T09:17:08.2775731Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0057s] [ 40%] 2025-07-17T09:17:08.2776256Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0052s] [ 40%] 2025-07-17T09:17:08.2776833Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0683s] [ 40%] 2025-07-17T09:17:08.2777342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0800s] [ 40%] 2025-07-17T09:17:08.2777802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0057s] [ 40%] 2025-07-17T09:17:08.2778275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0051s] [ 40%] 2025-07-17T09:17:08.2778736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0684s] [ 40%] 2025-07-17T09:17:08.2779200Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0068s] [ 40%] 2025-07-17T09:17:08.2779674Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0045s] [ 40%] 2025-07-17T09:17:08.2780146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0046s] [ 40%] 2025-07-17T09:17:08.2780613Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0083s] [ 40%] 2025-07-17T09:17:08.2781075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0155s] [ 40%] 2025-07-17T09:17:08.2781547Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0050s] [ 40%] 2025-07-17T09:17:08.2782014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0048s] [ 40%] 2025-07-17T09:17:08.2782485Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0077s] [ 40%] 2025-07-17T09:17:08.2782950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0065s] [ 40%] 2025-07-17T09:17:08.2783515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0045s] [ 40%] 2025-07-17T09:17:08.2784052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0042s] [ 40%] 2025-07-17T09:17:08.2784563Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0077s] [ 40%] 2025-07-17T09:17:08.2785028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0065s] [ 40%] 2025-07-17T09:17:08.2785491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0042s] [ 40%] 2025-07-17T09:17:08.2785960Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0043s] [ 40%] 2025-07-17T09:17:08.2786424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0076s] [ 40%] 2025-07-17T09:17:08.2786885Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0064s] [ 40%] 2025-07-17T09:17:08.2787491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2788082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2788667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2789256Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2789841Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2790426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2791167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2791816Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2792404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2792983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2793564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2794247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2794838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2795426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2796018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2796601Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2797189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2797778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2798553Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2799199Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2799793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2800381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2800971Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2801560Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2802141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2802726Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 40%] 2025-07-17T09:17:08.2803307Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2803894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2804482Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2805061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2805809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2806431Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2807024Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2807615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2808201Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2808785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2809375Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2809966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2810549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2811131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2811717Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2812292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2812992Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2813621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2814245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2814845Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2815427Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2816009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2816486Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0052s] [ 41%] 2025-07-17T09:17:08.2816950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 41%] 2025-07-17T09:17:08.2817419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0080s] [ 41%] 2025-07-17T09:17:08.2817877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0065s] [ 41%] 2025-07-17T09:17:08.2818354Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 41%] 2025-07-17T09:17:08.2818823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 41%] 2025-07-17T09:17:08.2819291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0080s] [ 41%] 2025-07-17T09:17:08.2819823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0230s] [ 41%] 2025-07-17T09:17:08.2820394Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0044s] [ 41%] 2025-07-17T09:17:08.2820908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0430s] [ 41%] 2025-07-17T09:17:08.2821367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0525s] [ 41%] 2025-07-17T09:17:08.2821833Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0488s] [ 41%] 2025-07-17T09:17:08.2822292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0045s] [ 41%] 2025-07-17T09:17:08.2822762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0042s] [ 41%] 2025-07-17T09:17:08.2823224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0415s] [ 41%] 2025-07-17T09:17:08.2823683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0553s] [ 41%] 2025-07-17T09:17:08.2824155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 41%] 2025-07-17T09:17:08.2824620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0119s] [ 41%] 2025-07-17T09:17:08.2825092Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0636s] [ 41%] 2025-07-17T09:17:08.2825557Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0306s] [ 41%] 2025-07-17T09:17:08.2826034Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0124s] [ 41%] 2025-07-17T09:17:08.2826505Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0297s] [ 41%] 2025-07-17T09:17:08.2827091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0634s] [ 41%] 2025-07-17T09:17:08.2827618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0183s] [ 41%] 2025-07-17T09:17:08.2828126Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0179s] [ 41%] 2025-07-17T09:17:08.2828589Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0260s] [ 41%] 2025-07-17T09:17:08.2829050Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0840s] [ 41%] 2025-07-17T09:17:08.2829515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0070s] [ 41%] 2025-07-17T09:17:08.2829981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0293s] [ 41%] 2025-07-17T09:17:08.2830441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0044s] [ 41%] 2025-07-17T09:17:08.2830905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0797s] [ 41%] 2025-07-17T09:17:08.2831361Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0193s] [ 41%] 2025-07-17T09:17:08.2831831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0077s] [ 41%] 2025-07-17T09:17:08.2832292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0053s] [ 41%] 2025-07-17T09:17:08.2832758Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0188s] [ 41%] 2025-07-17T09:17:08.2833230Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0076s] [ 41%] 2025-07-17T09:17:08.2833703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 41%] 2025-07-17T09:17:08.2834388Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0044s] [ 41%] 2025-07-17T09:17:08.2834920Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0083s] [ 41%] 2025-07-17T09:17:08.2835451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0066s] [ 41%] 2025-07-17T09:17:08.2835909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0042s] [ 41%] 2025-07-17T09:17:08.2836377Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0041s] [ 41%] 2025-07-17T09:17:08.2836838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0081s] [ 41%] 2025-07-17T09:17:08.2837296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0065s] [ 41%] 2025-07-17T09:17:08.2837764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0041s] [ 41%] 2025-07-17T09:17:08.2838229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0041s] [ 41%] 2025-07-17T09:17:08.2838694Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0082s] [ 41%] 2025-07-17T09:17:08.2839151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0065s] [ 41%] 2025-07-17T09:17:08.2839754Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2840346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2840925Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2841566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2842263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2842903Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2843491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2844074Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2844656Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2845232Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2845816Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2846400Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2846978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2847562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2848134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2848767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2849458Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2850090Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2850681Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2851263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2851860Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2852449Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2853037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2853625Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2854223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2854803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2855390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2856014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2856695Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2857340Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2857916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2858511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2859098Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2859691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2860288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2860866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2861459Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2862045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2862634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2863291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2863974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2864603Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2865185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2865770Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2866356Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2866937Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 41%] 2025-07-17T09:17:08.2867523Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2868099Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2868567Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0050s] [ 42%] 2025-07-17T09:17:08.2869032Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 42%] 2025-07-17T09:17:08.2869492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 42%] 2025-07-17T09:17:08.2869950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 42%] 2025-07-17T09:17:08.2870535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 42%] 2025-07-17T09:17:08.2871064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 42%] 2025-07-17T09:17:08.2871580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 42%] 2025-07-17T09:17:08.2872047Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 42%] 2025-07-17T09:17:08.2872518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 42%] 2025-07-17T09:17:08.2872975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 42%] 2025-07-17T09:17:08.2873429Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 42%] 2025-07-17T09:17:08.2873950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 42%] 2025-07-17T09:17:08.2874418Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 42%] 2025-07-17T09:17:08.2874888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 42%] 2025-07-17T09:17:08.2875346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 42%] 2025-07-17T09:17:08.2875800Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 42%] 2025-07-17T09:17:08.2876272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 42%] 2025-07-17T09:17:08.2876731Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 42%] 2025-07-17T09:17:08.2877193Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 42%] 2025-07-17T09:17:08.2877832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 42%] 2025-07-17T09:17:08.2878370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 42%] 2025-07-17T09:17:08.2878903Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 42%] 2025-07-17T09:17:08.2879363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 42%] 2025-07-17T09:17:08.2879827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 42%] 2025-07-17T09:17:08.2880285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0038s] [ 42%] 2025-07-17T09:17:08.2880744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 42%] 2025-07-17T09:17:08.2881196Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 42%] 2025-07-17T09:17:08.2881653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 42%] 2025-07-17T09:17:08.2882119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 42%] 2025-07-17T09:17:08.2882575Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 42%] 2025-07-17T09:17:08.2883036Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 42%] 2025-07-17T09:17:08.2883492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 42%] 2025-07-17T09:17:08.2883956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 42%] 2025-07-17T09:17:08.2884415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 42%] 2025-07-17T09:17:08.2884987Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 42%] 2025-07-17T09:17:08.2885526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 42%] 2025-07-17T09:17:08.2886040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 42%] 2025-07-17T09:17:08.2886506Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 42%] 2025-07-17T09:17:08.2886969Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 42%] 2025-07-17T09:17:08.2887439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 42%] 2025-07-17T09:17:08.2887899Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 42%] 2025-07-17T09:17:08.2888363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 42%] 2025-07-17T09:17:08.2888843Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 42%] 2025-07-17T09:17:08.2889296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 42%] 2025-07-17T09:17:08.2889760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 42%] 2025-07-17T09:17:08.2890216Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 42%] 2025-07-17T09:17:08.2890677Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 42%] 2025-07-17T09:17:08.2891130Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 42%] 2025-07-17T09:17:08.2891720Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2892411Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2893043Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2893686Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2894272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2894857Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2895437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2896013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2896591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2897163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2897744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2898321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2898895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2899577Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2900202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2900832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2901421Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2902003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2902581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2903159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2903745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2904326Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2904905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2905487Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2906061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2906765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2907402Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2908026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2908610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2909189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2909771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2910343Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2910924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2911506Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2912084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2912672Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2913259Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2913971Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2914688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2915320Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2915900Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2916479Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2917051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2917621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2918195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2918772Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2919346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2919925Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 42%] 2025-07-17T09:17:08.2920405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0097s] [ 42%] 2025-07-17T09:17:08.2920870Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0091s] [ 42%] 2025-07-17T09:17:08.2921453Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0277s] [ 42%] 2025-07-17T09:17:08.2921975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0187s] [ 42%] 2025-07-17T09:17:08.2922497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0092s] [ 42%] 2025-07-17T09:17:08.2922971Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0090s] [ 42%] 2025-07-17T09:17:08.2923441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0277s] [ 42%] 2025-07-17T09:17:08.2923911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0186s] [ 42%] 2025-07-17T09:17:08.2924378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0092s] [ 42%] 2025-07-17T09:17:08.2924845Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0091s] [ 42%] 2025-07-17T09:17:08.2925317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0277s] [ 42%] 2025-07-17T09:17:08.2925773Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0538s] [ 42%] 2025-07-17T09:17:08.2926241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0095s] [ 42%] 2025-07-17T09:17:08.2926702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0360s] [ 42%] 2025-07-17T09:17:08.2927166Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0863s] [ 42%] 2025-07-17T09:17:08.2927627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0767s] [ 42%] 2025-07-17T09:17:08.2928110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0238s] [ 42%] 2025-07-17T09:17:08.2928729Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0503s] [ 42%] 2025-07-17T09:17:08.2929251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.1989s] [ 43%] 2025-07-17T09:17:08.2929718Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.1186s] [ 43%] 2025-07-17T09:17:08.2930191Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0108s] [ 43%] 2025-07-17T09:17:08.2930662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0667s] [ 43%] 2025-07-17T09:17:08.2931131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.1680s] [ 43%] 2025-07-17T09:17:08.2931609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0828s] [ 43%] 2025-07-17T09:17:08.2932081Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0106s] [ 43%] 2025-07-17T09:17:08.2932542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0632s] [ 43%] 2025-07-17T09:17:08.2933010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.1492s] [ 43%] 2025-07-17T09:17:08.2933468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0971s] [ 43%] 2025-07-17T09:17:08.2933942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0476s] [ 43%] 2025-07-17T09:17:08.2934407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0475s] [ 43%] 2025-07-17T09:17:08.2934871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.1382s] [ 43%] 2025-07-17T09:17:08.2935390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0211s] [ 43%] 2025-07-17T09:17:08.2935962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0101s] [ 43%] 2025-07-17T09:17:08.2936485Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0101s] [ 43%] 2025-07-17T09:17:08.2936948Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0317s] [ 43%] 2025-07-17T09:17:08.2937420Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0208s] [ 43%] 2025-07-17T09:17:08.2937891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0101s] [ 43%] 2025-07-17T09:17:08.2938362Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0101s] [ 43%] 2025-07-17T09:17:08.2938839Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0419s] [ 43%] 2025-07-17T09:17:08.2939310Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0359s] [ 43%] 2025-07-17T09:17:08.2939776Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0103s] [ 43%] 2025-07-17T09:17:08.2940241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0101s] [ 43%] 2025-07-17T09:17:08.2940706Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0519s] [ 43%] 2025-07-17T09:17:08.2941169Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0210s] [ 43%] 2025-07-17T09:17:08.2941639Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0101s] [ 43%] 2025-07-17T09:17:08.2942105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0099s] [ 43%] 2025-07-17T09:17:08.2942691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0400s] [ 43%] 2025-07-17T09:17:08.2943209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0353s] [ 43%] 2025-07-17T09:17:08.2943851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2944447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2945037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2945616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2946204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2946790Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2947383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2947970Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2948559Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2949139Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2949771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2950447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2951089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2951670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2952251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2952827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2953420Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2954080Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2954670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2955256Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2955856Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2956451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2957135Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2957836Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2958483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2959068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2959653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2960231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2960811Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2961397Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2961978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2962565Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2963160Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2963745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2964384Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2965064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2965718Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2966310Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2966898Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2967496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2968083Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2968664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2969245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2969822Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2970416Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2971003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2971633Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2972314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 43%] 2025-07-17T09:17:08.2972834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0093s] [ 43%] 2025-07-17T09:17:08.2973305Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0089s] [ 43%] 2025-07-17T09:17:08.2973775Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0291s] [ 43%] 2025-07-17T09:17:08.2974251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0193s] [ 43%] 2025-07-17T09:17:08.2974728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0089s] [ 43%] 2025-07-17T09:17:08.2975196Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0088s] [ 43%] 2025-07-17T09:17:08.2975668Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0291s] [ 43%] 2025-07-17T09:17:08.2976131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0196s] [ 43%] 2025-07-17T09:17:08.2976597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0089s] [ 43%] 2025-07-17T09:17:08.2977063Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0088s] [ 43%] 2025-07-17T09:17:08.2977533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0291s] [ 43%] 2025-07-17T09:17:08.2977993Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0193s] [ 43%] 2025-07-17T09:17:08.2978457Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0089s] [ 43%] 2025-07-17T09:17:08.2979028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0088s] [ 43%] 2025-07-17T09:17:08.2979541Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0292s] [ 43%] 2025-07-17T09:17:08.2980052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0194s] [ 43%] 2025-07-17T09:17:08.2980521Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0103s] [ 43%] 2025-07-17T09:17:08.2980992Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0101s] [ 43%] 2025-07-17T09:17:08.2981465Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0339s] [ 43%] 2025-07-17T09:17:08.2981928Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0224s] [ 43%] 2025-07-17T09:17:08.2982406Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0102s] [ 43%] 2025-07-17T09:17:08.2982884Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0101s] [ 43%] 2025-07-17T09:17:08.2983361Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0339s] [ 43%] 2025-07-17T09:17:08.2983822Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0223s] [ 43%] 2025-07-17T09:17:08.2984292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0102s] [ 43%] 2025-07-17T09:17:08.2984760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0102s] [ 43%] 2025-07-17T09:17:08.2985227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0340s] [ 43%] 2025-07-17T09:17:08.2985747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0229s] [ 43%] 2025-07-17T09:17:08.2986311Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0107s] [ 43%] 2025-07-17T09:17:08.2986833Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0102s] [ 43%] 2025-07-17T09:17:08.2987300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0342s] [ 43%] 2025-07-17T09:17:08.2987767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0224s] [ 43%] 2025-07-17T09:17:08.2988244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0102s] [ 43%] 2025-07-17T09:17:08.2988710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0101s] [ 43%] 2025-07-17T09:17:08.2989178Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0340s] [ 43%] 2025-07-17T09:17:08.2989642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0223s] [ 43%] 2025-07-17T09:17:08.2990120Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0103s] [ 43%] 2025-07-17T09:17:08.2990589Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0102s] [ 43%] 2025-07-17T09:17:08.2991059Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0340s] [ 44%] 2025-07-17T09:17:08.2991541Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0224s] [ 44%] 2025-07-17T09:17:08.2992004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0102s] [ 44%] 2025-07-17T09:17:08.2992468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0101s] [ 44%] 2025-07-17T09:17:08.2993118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0342s] [ 44%] 2025-07-17T09:17:08.2993703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0224s] [ 44%] 2025-07-17T09:17:08.2994277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0102s] [ 44%] 2025-07-17T09:17:08.2994746Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0102s] [ 44%] 2025-07-17T09:17:08.2995221Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0341s] [ 44%] 2025-07-17T09:17:08.2995687Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0223s] [ 44%] 2025-07-17T09:17:08.2996285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.2996881Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.2997494Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.2998079Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.2998668Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.2999267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.2999854Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3000506Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3001216Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3001855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3002440Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3003018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3003602Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3004188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3004767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3005345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3005936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3006527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3007116Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3007751Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3008451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3009092Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3009685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3010273Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3010854Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3011437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3012016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3012595Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3013185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3013764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3014347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3014981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3015665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3016308Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3016895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3017482Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3018075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3018660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3019252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3019838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3020427Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3021008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3021585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3022227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3022930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3023571Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3024158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3024736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3025206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0057s] [ 44%] 2025-07-17T09:17:08.3025670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0056s] [ 44%] 2025-07-17T09:17:08.3026137Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0140s] [ 44%] 2025-07-17T09:17:08.3026597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0096s] [ 44%] 2025-07-17T09:17:08.3027068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0055s] [ 44%] 2025-07-17T09:17:08.3027541Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0055s] [ 44%] 2025-07-17T09:17:08.3028001Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0128s] [ 44%] 2025-07-17T09:17:08.3028468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0096s] [ 44%] 2025-07-17T09:17:08.3028927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0054s] [ 44%] 2025-07-17T09:17:08.3029494Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0056s] [ 44%] 2025-07-17T09:17:08.3030003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0134s] [ 44%] 2025-07-17T09:17:08.3030518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0094s] [ 44%] 2025-07-17T09:17:08.3030988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0048s] [ 44%] 2025-07-17T09:17:08.3031454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0048s] [ 44%] 2025-07-17T09:17:08.3031918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0121s] [ 44%] 2025-07-17T09:17:08.3032378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0089s] [ 44%] 2025-07-17T09:17:08.3032850Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0061s] [ 44%] 2025-07-17T09:17:08.3033315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0063s] [ 44%] 2025-07-17T09:17:08.3033784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0175s] [ 44%] 2025-07-17T09:17:08.3034325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0129s] [ 44%] 2025-07-17T09:17:08.3034796Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0064s] [ 44%] 2025-07-17T09:17:08.3035266Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0062s] [ 44%] 2025-07-17T09:17:08.3035732Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0176s] [ 44%] 2025-07-17T09:17:08.3036199Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0132s] [ 44%] 2025-07-17T09:17:08.3036789Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0065s] [ 44%] 2025-07-17T09:17:08.3037315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0064s] [ 44%] 2025-07-17T09:17:08.3037834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0176s] [ 44%] 2025-07-17T09:17:08.3038288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0130s] [ 44%] 2025-07-17T09:17:08.3038759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0063s] [ 44%] 2025-07-17T09:17:08.3039220Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0062s] [ 44%] 2025-07-17T09:17:08.3039683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0173s] [ 44%] 2025-07-17T09:17:08.3040138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0126s] [ 44%] 2025-07-17T09:17:08.3040607Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0062s] [ 44%] 2025-07-17T09:17:08.3041072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0062s] [ 44%] 2025-07-17T09:17:08.3041535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0176s] [ 44%] 2025-07-17T09:17:08.3041999Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0132s] [ 44%] 2025-07-17T09:17:08.3042467Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0064s] [ 44%] 2025-07-17T09:17:08.3042934Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0063s] [ 44%] 2025-07-17T09:17:08.3043397Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0174s] [ 44%] 2025-07-17T09:17:08.3043974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0128s] [ 44%] 2025-07-17T09:17:08.3044486Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0060s] [ 44%] 2025-07-17T09:17:08.3045008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0060s] [ 44%] 2025-07-17T09:17:08.3045467Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0173s] [ 44%] 2025-07-17T09:17:08.3045926Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0127s] [ 44%] 2025-07-17T09:17:08.3046394Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0061s] [ 44%] 2025-07-17T09:17:08.3046855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0061s] [ 44%] 2025-07-17T09:17:08.3047324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0171s] [ 44%] 2025-07-17T09:17:08.3047786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0126s] [ 44%] 2025-07-17T09:17:08.3048374Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3048964Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3049553Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3050138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3050816Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3051502Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3052146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3052734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3053331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3053916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 44%] 2025-07-17T09:17:08.3054500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3055079Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3055654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3056237Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3056819Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3057395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3057988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3058727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3059370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3059953Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3060541Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3061132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3061715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3062309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3062899Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3063474Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3064058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3064630Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3065215Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3065961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3066598Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3067185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3067775Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3068366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3068950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3069537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3070127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3070709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3071300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3071888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3072467Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3073209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3073861Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3074517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3075106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3075685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3076267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3076854Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3077330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0053s] [ 45%] 2025-07-17T09:17:08.3077801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3078263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 45%] 2025-07-17T09:17:08.3078729Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 45%] 2025-07-17T09:17:08.3079191Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3079656Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3080254Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 45%] 2025-07-17T09:17:08.3080778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 45%] 2025-07-17T09:17:08.3081303Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 45%] 2025-07-17T09:17:08.3081760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 45%] 2025-07-17T09:17:08.3082226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 45%] 2025-07-17T09:17:08.3082677Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 45%] 2025-07-17T09:17:08.3083138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 45%] 2025-07-17T09:17:08.3083600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 45%] 2025-07-17T09:17:08.3084060Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 45%] 2025-07-17T09:17:08.3084518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3084986Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 45%] 2025-07-17T09:17:08.3085461Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 45%] 2025-07-17T09:17:08.3085926Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 45%] 2025-07-17T09:17:08.3086391Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 45%] 2025-07-17T09:17:08.3086864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 45%] 2025-07-17T09:17:08.3087440Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3087964Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 45%] 2025-07-17T09:17:08.3088473Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 45%] 2025-07-17T09:17:08.3088937Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3089401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0038s] [ 45%] 2025-07-17T09:17:08.3089859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 45%] 2025-07-17T09:17:08.3090314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0039s] [ 45%] 2025-07-17T09:17:08.3090774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 45%] 2025-07-17T09:17:08.3091235Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3091692Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0039s] [ 45%] 2025-07-17T09:17:08.3092150Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 45%] 2025-07-17T09:17:08.3092617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3093082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3093546Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 45%] 2025-07-17T09:17:08.3094004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 45%] 2025-07-17T09:17:08.3094586Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3095101Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 45%] 2025-07-17T09:17:08.3095617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 45%] 2025-07-17T09:17:08.3096076Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 45%] 2025-07-17T09:17:08.3096544Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3097004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3097460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0041s] [ 45%] 2025-07-17T09:17:08.3097921Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0040s] [ 45%] 2025-07-17T09:17:08.3098377Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3098840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0037s] [ 45%] 2025-07-17T09:17:08.3099297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0040s] [ 45%] 2025-07-17T09:17:08.3099758Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0039s] [ 45%] 2025-07-17T09:17:08.3100350Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3100933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3101622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3102273Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3103036Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3103625Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3104209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3104793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3105373Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3105951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3106532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3107105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3107688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3108270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3108925Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3109620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3110258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3110851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3111430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3112017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3112615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3113203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3113786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3114434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3115020Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3115611Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3116257Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3116967Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3117619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3118203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 45%] 2025-07-17T09:17:08.3118787Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3119364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3119954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3120537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3121122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3121701Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3122290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3122876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3123538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3124227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3124860Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3125435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3126015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3126586Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3127174Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3127758Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3128335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3128919Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3129384Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 46%] 2025-07-17T09:17:08.3129852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 46%] 2025-07-17T09:17:08.3130314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 46%] 2025-07-17T09:17:08.3130914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 46%] 2025-07-17T09:17:08.3131438Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 46%] 2025-07-17T09:17:08.3131954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 46%] 2025-07-17T09:17:08.3132414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 46%] 2025-07-17T09:17:08.3132876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 46%] 2025-07-17T09:17:08.3133342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 46%] 2025-07-17T09:17:08.3133805Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 46%] 2025-07-17T09:17:08.3134259Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 46%] 2025-07-17T09:17:08.3134719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 46%] 2025-07-17T09:17:08.3135180Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 46%] 2025-07-17T09:17:08.3135642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 46%] 2025-07-17T09:17:08.3136102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 46%] 2025-07-17T09:17:08.3136562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 46%] 2025-07-17T09:17:08.3137035Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 46%] 2025-07-17T09:17:08.3137497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 46%] 2025-07-17T09:17:08.3138073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 46%] 2025-07-17T09:17:08.3138585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0038s] [ 46%] 2025-07-17T09:17:08.3139117Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 46%] 2025-07-17T09:17:08.3139582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 46%] 2025-07-17T09:17:08.3140052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 46%] 2025-07-17T09:17:08.3140518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 46%] 2025-07-17T09:17:08.3140974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 46%] 2025-07-17T09:17:08.3141439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 46%] 2025-07-17T09:17:08.3141895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0038s] [ 46%] 2025-07-17T09:17:08.3142357Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 46%] 2025-07-17T09:17:08.3142815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 46%] 2025-07-17T09:17:08.3143279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 46%] 2025-07-17T09:17:08.3143741Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 46%] 2025-07-17T09:17:08.3144200Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 46%] 2025-07-17T09:17:08.3144670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 46%] 2025-07-17T09:17:08.3145241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 46%] 2025-07-17T09:17:08.3145762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 46%] 2025-07-17T09:17:08.3146277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 46%] 2025-07-17T09:17:08.3146747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 46%] 2025-07-17T09:17:08.3147219Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 46%] 2025-07-17T09:17:08.3147685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 46%] 2025-07-17T09:17:08.3148152Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 46%] 2025-07-17T09:17:08.3148610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 46%] 2025-07-17T09:17:08.3149074Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 46%] 2025-07-17T09:17:08.3149528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 46%] 2025-07-17T09:17:08.3149993Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 46%] 2025-07-17T09:17:08.3150462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 46%] 2025-07-17T09:17:08.3150922Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 46%] 2025-07-17T09:17:08.3151382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 46%] 2025-07-17T09:17:08.3151835Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0081s] [ 46%] 2025-07-17T09:17:08.3152543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0014s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3153179Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3153816Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3154464Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3155049Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3155638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3156227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3156809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3157393Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3157975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3158557Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3159127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3159878Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3160523Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3161155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3161738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3162321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3162908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3163495Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3164077Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3164664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3165249Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3165836Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3166423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3167110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3167753Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3168377Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3168961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3169552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3170130Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3170712Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3171284Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3171873Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3172465Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3173048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3173631Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3174337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3174984Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3175620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3176207Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0013s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3176790Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3177374Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3177952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3178533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3179112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3179698Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3180278Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3180859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 46%] 2025-07-17T09:17:08.3181442Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 46%] 2025-07-17T09:17:08.3181949Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 46%] 2025-07-17T09:17:08.3182478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0058s] [ 47%] 2025-07-17T09:17:08.3182931Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0112s] [ 47%] 2025-07-17T09:17:08.3183401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0047s] [ 47%] 2025-07-17T09:17:08.3183866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3184324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0120s] [ 47%] 2025-07-17T09:17:08.3184792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0050s] [ 47%] 2025-07-17T09:17:08.3185246Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0218s] [ 47%] 2025-07-17T09:17:08.3185703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0053s] [ 47%] 2025-07-17T09:17:08.3186157Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0088s] [ 47%] 2025-07-17T09:17:08.3186608Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0054s] [ 47%] 2025-07-17T09:17:08.3187070Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0041s] [ 47%] 2025-07-17T09:17:08.3187522Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 47%] 2025-07-17T09:17:08.3187977Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 47%] 2025-07-17T09:17:08.3188533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 47%] 2025-07-17T09:17:08.3189102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 47%] 2025-07-17T09:17:08.3189609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3190070Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 47%] 2025-07-17T09:17:08.3190530Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3190995Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 47%] 2025-07-17T09:17:08.3191461Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 47%] 2025-07-17T09:17:08.3191918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3192389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3192847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 47%] 2025-07-17T09:17:08.3193306Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 47%] 2025-07-17T09:17:08.3193762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3194274Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3194738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 47%] 2025-07-17T09:17:08.3195194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 47%] 2025-07-17T09:17:08.3195719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0036s] [ 47%] 2025-07-17T09:17:08.3196293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 47%] 2025-07-17T09:17:08.3196826Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 47%] 2025-07-17T09:17:08.3197283Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 47%] 2025-07-17T09:17:08.3197747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 47%] 2025-07-17T09:17:08.3198208Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3198671Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 47%] 2025-07-17T09:17:08.3199132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 47%] 2025-07-17T09:17:08.3199590Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3200056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3200513Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 47%] 2025-07-17T09:17:08.3200967Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 47%] 2025-07-17T09:17:08.3201426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 47%] 2025-07-17T09:17:08.3201878Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3202337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 47%] 2025-07-17T09:17:08.3202845Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 47%] 2025-07-17T09:17:08.3203413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3203923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 47%] 2025-07-17T09:17:08.3204516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3205098Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3205674Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3206253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3206833Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3207418Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3208004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3208582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3209163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3209736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3210425Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3211072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3211707Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3212284Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3212852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3213426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3214013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3214596Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3215180Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3215766Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3216354Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3216939Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3217629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3218387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3219008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3219592Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3220170Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3220740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3221323Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3221895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3222481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3223062Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3223643Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3224226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3224856Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3225532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3226167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3226751Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3227336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3227914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3228489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3229069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3229639Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3230216Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3230797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3231370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3231996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3232665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_127_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 47%] 2025-07-17T09:17:08.3233191Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0050s] [ 47%] 2025-07-17T09:17:08.3233652Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 47%] 2025-07-17T09:17:08.3234195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0110s] [ 47%] 2025-07-17T09:17:08.3234669Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0073s] [ 47%] 2025-07-17T09:17:08.3235141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 47%] 2025-07-17T09:17:08.3235611Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 47%] 2025-07-17T09:17:08.3236079Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0110s] [ 47%] 2025-07-17T09:17:08.3236554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0073s] [ 47%] 2025-07-17T09:17:08.3237016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 47%] 2025-07-17T09:17:08.3237478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0040s] [ 47%] 2025-07-17T09:17:08.3237945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0111s] [ 47%] 2025-07-17T09:17:08.3238402Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0073s] [ 47%] 2025-07-17T09:17:08.3238868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 47%] 2025-07-17T09:17:08.3239464Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0040s] [ 47%] 2025-07-17T09:17:08.3239991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0109s] [ 47%] 2025-07-17T09:17:08.3240512Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0073s] [ 47%] 2025-07-17T09:17:08.3240979Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0044s] [ 47%] 2025-07-17T09:17:08.3241450Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0045s] [ 47%] 2025-07-17T09:17:08.3241912Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0119s] [ 47%] 2025-07-17T09:17:08.3242383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0079s] [ 47%] 2025-07-17T09:17:08.3242854Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0044s] [ 47%] 2025-07-17T09:17:08.3243331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0043s] [ 47%] 2025-07-17T09:17:08.3243800Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0118s] [ 48%] 2025-07-17T09:17:08.3244267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0079s] [ 48%] 2025-07-17T09:17:08.3244735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0043s] [ 48%] 2025-07-17T09:17:08.3245197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0043s] [ 48%] 2025-07-17T09:17:08.3245660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0117s] [ 48%] 2025-07-17T09:17:08.3246120Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0078s] [ 48%] 2025-07-17T09:17:08.3246700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0043s] [ 48%] 2025-07-17T09:17:08.3247213Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0043s] [ 48%] 2025-07-17T09:17:08.3247730Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0117s] [ 48%] 2025-07-17T09:17:08.3248194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0078s] [ 48%] 2025-07-17T09:17:08.3248662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0043s] [ 48%] 2025-07-17T09:17:08.3249134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0043s] [ 48%] 2025-07-17T09:17:08.3249600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0119s] [ 48%] 2025-07-17T09:17:08.3250068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0086s] [ 48%] 2025-07-17T09:17:08.3250540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0053s] [ 48%] 2025-07-17T09:17:08.3251012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0044s] [ 48%] 2025-07-17T09:17:08.3251483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0119s] [ 48%] 2025-07-17T09:17:08.3251951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0080s] [ 48%] 2025-07-17T09:17:08.3252418Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0044s] [ 48%] 2025-07-17T09:17:08.3252877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0044s] [ 48%] 2025-07-17T09:17:08.3253341Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0126s] [ 48%] 2025-07-17T09:17:08.3253914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0080s] [ 48%] 2025-07-17T09:17:08.3254434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0044s] [ 48%] 2025-07-17T09:17:08.3254948Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0044s] [ 48%] 2025-07-17T09:17:08.3255409Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0120s] [ 48%] 2025-07-17T09:17:08.3255879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0080s] [ 48%] 2025-07-17T09:17:08.3256465Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3257051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3257647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3258227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3258817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3259400Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3259994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3260579Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3261318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3261958Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3262534Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3263117Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3263700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3264279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3264862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3265451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3266050Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3266642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3267226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3267815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3268525Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3269176Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3269815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3270400Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3270984Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3271563Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3272151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3272735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3273316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3273964Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3274555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3275140Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3275864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3276513Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3277170Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3277754Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3278346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3278936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3279520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3280109Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3280692Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3281279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3281862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3282437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3283132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3283768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3284406Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3284999Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 48%] 2025-07-17T09:17:08.3285473Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0055s] [ 48%] 2025-07-17T09:17:08.3285944Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 48%] 2025-07-17T09:17:08.3286408Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0126s] [ 48%] 2025-07-17T09:17:08.3286875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0083s] [ 48%] 2025-07-17T09:17:08.3287353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 48%] 2025-07-17T09:17:08.3287821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 48%] 2025-07-17T09:17:08.3288294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0126s] [ 48%] 2025-07-17T09:17:08.3288760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0085s] [ 48%] 2025-07-17T09:17:08.3289225Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0043s] [ 48%] 2025-07-17T09:17:08.3289684Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0042s] [ 48%] 2025-07-17T09:17:08.3290268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0125s] [ 48%] 2025-07-17T09:17:08.3290786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0083s] [ 48%] 2025-07-17T09:17:08.3291317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0042s] [ 48%] 2025-07-17T09:17:08.3291784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0041s] [ 48%] 2025-07-17T09:17:08.3292247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0126s] [ 48%] 2025-07-17T09:17:08.3292709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0083s] [ 48%] 2025-07-17T09:17:08.3293177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 48%] 2025-07-17T09:17:08.3293660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0045s] [ 48%] 2025-07-17T09:17:08.3294129Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0136s] [ 48%] 2025-07-17T09:17:08.3294595Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0090s] [ 48%] 2025-07-17T09:17:08.3295067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 48%] 2025-07-17T09:17:08.3295535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0045s] [ 48%] 2025-07-17T09:17:08.3296011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0136s] [ 48%] 2025-07-17T09:17:08.3296474Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0090s] [ 48%] 2025-07-17T09:17:08.3296944Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0045s] [ 48%] 2025-07-17T09:17:08.3297526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0046s] [ 48%] 2025-07-17T09:17:08.3298041Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0135s] [ 48%] 2025-07-17T09:17:08.3298555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0089s] [ 48%] 2025-07-17T09:17:08.3299023Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0046s] [ 48%] 2025-07-17T09:17:08.3299496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0045s] [ 48%] 2025-07-17T09:17:08.3299963Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0135s] [ 48%] 2025-07-17T09:17:08.3300436Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0089s] [ 48%] 2025-07-17T09:17:08.3300907Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0045s] [ 48%] 2025-07-17T09:17:08.3301378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0045s] [ 48%] 2025-07-17T09:17:08.3301849Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0135s] [ 48%] 2025-07-17T09:17:08.3302309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0089s] [ 48%] 2025-07-17T09:17:08.3302784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 48%] 2025-07-17T09:17:08.3303249Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0045s] [ 48%] 2025-07-17T09:17:08.3303731Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0135s] [ 48%] 2025-07-17T09:17:08.3304201Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0089s] [ 48%] 2025-07-17T09:17:08.3304775Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0045s] [ 48%] 2025-07-17T09:17:08.3305302Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0046s] [ 48%] 2025-07-17T09:17:08.3305808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0135s] [ 49%] 2025-07-17T09:17:08.3306273Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0089s] [ 49%] 2025-07-17T09:17:08.3306734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0045s] [ 49%] 2025-07-17T09:17:08.3307198Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0045s] [ 49%] 2025-07-17T09:17:08.3307666Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0136s] [ 49%] 2025-07-17T09:17:08.3308130Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0089s] [ 49%] 2025-07-17T09:17:08.3308730Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3309315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3309902Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3310482Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3311075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3311727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3312411Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3313051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3313638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3314283Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3314869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3315448Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3316037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3316621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3317202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3317785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3318372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3319025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3319746Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3320385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3320981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3321571Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3322158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3322747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3323333Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3323917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3324495Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3325078Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3325662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3326328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3327024Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3327648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3328246Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3328836Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3329418Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3330009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3330597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3331197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3331789Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3332370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3332956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3333592Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3334394Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3335034Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3335614Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3336204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3336782Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3337365Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3337839Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0047s] [ 49%] 2025-07-17T09:17:08.3338302Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3338773Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 49%] 2025-07-17T09:17:08.3339234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 49%] 2025-07-17T09:17:08.3339708Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 49%] 2025-07-17T09:17:08.3340171Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 49%] 2025-07-17T09:17:08.3340639Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 49%] 2025-07-17T09:17:08.3341221Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 49%] 2025-07-17T09:17:08.3341735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 49%] 2025-07-17T09:17:08.3342251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0039s] [ 49%] 2025-07-17T09:17:08.3342704Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 49%] 2025-07-17T09:17:08.3343168Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3343625Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 49%] 2025-07-17T09:17:08.3344086Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 49%] 2025-07-17T09:17:08.3344548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 49%] 2025-07-17T09:17:08.3345008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3345480Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 49%] 2025-07-17T09:17:08.3345938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 49%] 2025-07-17T09:17:08.3346414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 49%] 2025-07-17T09:17:08.3346871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 49%] 2025-07-17T09:17:08.3347343Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 49%] 2025-07-17T09:17:08.3347809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 49%] 2025-07-17T09:17:08.3348400Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 49%] 2025-07-17T09:17:08.3348918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 49%] 2025-07-17T09:17:08.3349424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 49%] 2025-07-17T09:17:08.3349893Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 49%] 2025-07-17T09:17:08.3350351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3350809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3351269Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 49%] 2025-07-17T09:17:08.3351728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 49%] 2025-07-17T09:17:08.3352188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3352641Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3353108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 49%] 2025-07-17T09:17:08.3353574Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 49%] 2025-07-17T09:17:08.3354112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3354576Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3355035Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 49%] 2025-07-17T09:17:08.3355649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 49%] 2025-07-17T09:17:08.3356176Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 49%] 2025-07-17T09:17:08.3356712Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3357167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 49%] 2025-07-17T09:17:08.3357636Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 49%] 2025-07-17T09:17:08.3358097Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3358546Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3359009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 49%] 2025-07-17T09:17:08.3359469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 49%] 2025-07-17T09:17:08.3359930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3360382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 49%] 2025-07-17T09:17:08.3360979Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3361573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3362159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3362858Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3363488Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3364125Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3364713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3365290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3365875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3366453Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3367032Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3367612Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3368189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 49%] 2025-07-17T09:17:08.3368771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3369350Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3370036Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3370675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3371307Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3371898Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3372477Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3373070Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3373661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3374240Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3374823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3375404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3375988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3376568Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3377210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3377893Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3378521Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3379112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3379693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3380277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3380866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3381449Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3382031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3382620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3383204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3383790Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3384418Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3385110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3385753Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3386330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3386919Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3387502Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3388084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3388678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3389252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 50%] 2025-07-17T09:17:08.3389721Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0041s] [ 50%] 2025-07-17T09:17:08.3390182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 50%] 2025-07-17T09:17:08.3390643Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3391104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3391616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 50%] 2025-07-17T09:17:08.3392183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 50%] 2025-07-17T09:17:08.3392699Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3393160Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 50%] 2025-07-17T09:17:08.3393616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 50%] 2025-07-17T09:17:08.3394185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 50%] 2025-07-17T09:17:08.3394656Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3395130Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3395593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 50%] 2025-07-17T09:17:08.3396069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 50%] 2025-07-17T09:17:08.3396575Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3402048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3402584Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3403060Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3403526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 50%] 2025-07-17T09:17:08.3404108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3404707Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3405253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3405722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 50%] 2025-07-17T09:17:08.3406185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 50%] 2025-07-17T09:17:08.3406649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3407098Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3407568Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 50%] 2025-07-17T09:17:08.3408027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3408490Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 50%] 2025-07-17T09:17:08.3408951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3409413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 50%] 2025-07-17T09:17:08.3409880Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 50%] 2025-07-17T09:17:08.3410347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3410812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3411327Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 50%] 2025-07-17T09:17:08.3411889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3412405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3412865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3413339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 50%] 2025-07-17T09:17:08.3413798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3414262Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3414723Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3415178Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3415639Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3416093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3416552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3417009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3417470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3417931Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3418388Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 50%] 2025-07-17T09:17:08.3419006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3419508Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 50%] 2025-07-17T09:17:08.3419968Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 50%] 2025-07-17T09:17:08.3420425Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 50%] 2025-07-17T09:17:08.3420899Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3421360Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 50%] 2025-07-17T09:17:08.3421810Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3422272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 50%] 2025-07-17T09:17:08.3422720Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 50%] 2025-07-17T09:17:08.3423172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3423620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 50%] 2025-07-17T09:17:08.3424075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 50%] 2025-07-17T09:17:08.3424534Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3424986Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 50%] 2025-07-17T09:17:08.3425448Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 50%] 2025-07-17T09:17:08.3426016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3426528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3427027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 50%] 2025-07-17T09:17:08.3427494Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3427956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3428422Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3428881Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 50%] 2025-07-17T09:17:08.3429340Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3429800Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3430252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 50%] 2025-07-17T09:17:08.3430710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 50%] 2025-07-17T09:17:08.3431161Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3431617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 50%] 2025-07-17T09:17:08.3432078Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 50%] 2025-07-17T09:17:08.3432528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 50%] 2025-07-17T09:17:08.3433113Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 50%] 2025-07-17T09:17:08.3433643Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3434234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 51%] 2025-07-17T09:17:08.3434689Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 51%] 2025-07-17T09:17:08.3435159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3435621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3436076Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 51%] 2025-07-17T09:17:08.3436538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 51%] 2025-07-17T09:17:08.3436996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3437452Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3437897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 51%] 2025-07-17T09:17:08.3438352Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 51%] 2025-07-17T09:17:08.3438811Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3439263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3439715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 51%] 2025-07-17T09:17:08.3440291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 51%] 2025-07-17T09:17:08.3440817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 51%] 2025-07-17T09:17:08.3441334Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0026s] [ 51%] 2025-07-17T09:17:08.3441792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3442255Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3442719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0027s] [ 51%] 2025-07-17T09:17:08.3443185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0027s] [ 51%] 2025-07-17T09:17:08.3443648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 51%] 2025-07-17T09:17:08.3444111Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 51%] 2025-07-17T09:17:08.3444562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0027s] [ 51%] 2025-07-17T09:17:08.3445017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 51%] 2025-07-17T09:17:08.3445472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3445923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0029s] [ 51%] 2025-07-17T09:17:08.3446381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0028s] [ 51%] 2025-07-17T09:17:08.3446829Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 51%] 2025-07-17T09:17:08.3447404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 51%] 2025-07-17T09:17:08.3447911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3448426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 51%] 2025-07-17T09:17:08.3448888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3449349Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3449809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3450269Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3450739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3451206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3451677Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3452138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3452594Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3453051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3453500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3453961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3454647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3455161Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3455672Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3456133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 51%] 2025-07-17T09:17:08.3456599Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3457056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3457517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3457983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3458452Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 51%] 2025-07-17T09:17:08.3458919Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3459374Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3459838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 51%] 2025-07-17T09:17:08.3460295Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 51%] 2025-07-17T09:17:08.3460756Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 51%] 2025-07-17T09:17:08.3461209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 51%] 2025-07-17T09:17:08.3461799Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 51%] 2025-07-17T09:17:08.3462307Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 51%] 2025-07-17T09:17:08.3462818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 51%] 2025-07-17T09:17:08.3463279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0031s] [ 51%] 2025-07-17T09:17:08.3463744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 51%] 2025-07-17T09:17:08.3464206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 51%] 2025-07-17T09:17:08.3464660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 51%] 2025-07-17T09:17:08.3465116Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3465579Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0028s] [ 51%] 2025-07-17T09:17:08.3466040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 51%] 2025-07-17T09:17:08.3466497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3466954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3467409Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0028s] [ 51%] 2025-07-17T09:17:08.3467860Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 51%] 2025-07-17T09:17:08.3468316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3468872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3469385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0028s] [ 51%] 2025-07-17T09:17:08.3469893Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 51%] 2025-07-17T09:17:08.3470344Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3470802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3471264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3471727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 51%] 2025-07-17T09:17:08.3472177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3472637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3473104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3473560Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 51%] 2025-07-17T09:17:08.3474106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3474566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3475026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3475476Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 51%] 2025-07-17T09:17:08.3476061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3476581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3477098Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 51%] 2025-07-17T09:17:08.3477556Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3478010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3478470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3478927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3479388Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3479846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3480300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3480769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3481223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3481686Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3482148Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3482606Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3483111Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3483662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3484163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3484614Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0031s] [ 51%] 2025-07-17T09:17:08.3485074Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3485527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3485989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0033s] [ 51%] 2025-07-17T09:17:08.3486453Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 51%] 2025-07-17T09:17:08.3486909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0027s] [ 51%] 2025-07-17T09:17:08.3487366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 51%] 2025-07-17T09:17:08.3487818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 51%] 2025-07-17T09:17:08.3488279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0028s] [ 51%] 2025-07-17T09:17:08.3488737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0027s] [ 52%] 2025-07-17T09:17:08.3489198Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3489659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3490183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0027s] [ 52%] 2025-07-17T09:17:08.3490744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 52%] 2025-07-17T09:17:08.3491245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3491696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3492159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0027s] [ 52%] 2025-07-17T09:17:08.3492617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0027s] [ 52%] 2025-07-17T09:17:08.3493064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3493518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3493983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3494440Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3494901Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 52%] 2025-07-17T09:17:08.3495349Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 52%] 2025-07-17T09:17:08.3495817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3496274Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3496735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3497191Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3497797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3498304Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3498753Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 52%] 2025-07-17T09:17:08.3499209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0039s] [ 52%] 2025-07-17T09:17:08.3499661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3500121Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3500573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3501025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3501492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3501948Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3502405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0031s] [ 52%] 2025-07-17T09:17:08.3502858Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3503320Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3503779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3504229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3504818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3505332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3505840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3506286Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3506745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3507205Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3507654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3508110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3508563Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3509019Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3509468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3509922Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3510377Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0029s] [ 52%] 2025-07-17T09:17:08.3510833Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3511287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3511850Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3512357Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3512870Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3513324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3513769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3514288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3514744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3515196Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0028s] [ 52%] 2025-07-17T09:17:08.3515649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3516094Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3516554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 52%] 2025-07-17T09:17:08.3517011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3517502Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 52%] 2025-07-17T09:17:08.3517981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 52%] 2025-07-17T09:17:08.3518435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3519036Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3519549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 52%] 2025-07-17T09:17:08.3520064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 52%] 2025-07-17T09:17:08.3520510Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3520966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3521421Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 52%] 2025-07-17T09:17:08.3521865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 52%] 2025-07-17T09:17:08.3522321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3522768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0031s] [ 52%] 2025-07-17T09:17:08.3523218Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0033s] [ 52%] 2025-07-17T09:17:08.3523663Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 52%] 2025-07-17T09:17:08.3524123Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3524584Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3525038Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 52%] 2025-07-17T09:17:08.3525492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0032s] [ 52%] 2025-07-17T09:17:08.3526006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3526566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3527068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0032s] [ 52%] 2025-07-17T09:17:08.3527527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0031s] [ 52%] 2025-07-17T09:17:08.3527981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3528433Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3528886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 52%] 2025-07-17T09:17:08.3529330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 52%] 2025-07-17T09:17:08.3529787Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3530237Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0030s] [ 52%] 2025-07-17T09:17:08.3530693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0031s] [ 52%] 2025-07-17T09:17:08.3531142Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_4_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0032s] [ 52%] 2025-07-17T09:17:08.3531610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 52%] 2025-07-17T09:17:08.3532078Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 52%] 2025-07-17T09:17:08.3532539Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0043s] [ 52%] 2025-07-17T09:17:08.3533003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 52%] 2025-07-17T09:17:08.3533624Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 52%] 2025-07-17T09:17:08.3534138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 52%] 2025-07-17T09:17:08.3534604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 52%] 2025-07-17T09:17:08.3535066Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 52%] 2025-07-17T09:17:08.3535526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 52%] 2025-07-17T09:17:08.3535982Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 52%] 2025-07-17T09:17:08.3536443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0042s] [ 52%] 2025-07-17T09:17:08.3536896Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 52%] 2025-07-17T09:17:08.3537364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 52%] 2025-07-17T09:17:08.3537819Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 52%] 2025-07-17T09:17:08.3538282Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0043s] [ 52%] 2025-07-17T09:17:08.3538742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0038s] [ 52%] 2025-07-17T09:17:08.3539204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 52%] 2025-07-17T09:17:08.3539671Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 52%] 2025-07-17T09:17:08.3540192Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0045s] [ 52%] 2025-07-17T09:17:08.3540776Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 52%] 2025-07-17T09:17:08.3541292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 52%] 2025-07-17T09:17:08.3541761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 52%] 2025-07-17T09:17:08.3542234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0045s] [ 52%] 2025-07-17T09:17:08.3542697Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 52%] 2025-07-17T09:17:08.3543163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 52%] 2025-07-17T09:17:08.3543616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0036s] [ 53%] 2025-07-17T09:17:08.3544088Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 53%] 2025-07-17T09:17:08.3544545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0040s] [ 53%] 2025-07-17T09:17:08.3545008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0035s] [ 53%] 2025-07-17T09:17:08.3545473Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0035s] [ 53%] 2025-07-17T09:17:08.3545930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0045s] [ 53%] 2025-07-17T09:17:08.3546394Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0040s] [ 53%] 2025-07-17T09:17:08.3546862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 53%] 2025-07-17T09:17:08.3547387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 53%] 2025-07-17T09:17:08.3547963Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0045s] [ 53%] 2025-07-17T09:17:08.3548490Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 53%] 2025-07-17T09:17:08.3548957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0036s] [ 53%] 2025-07-17T09:17:08.3549423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 53%] 2025-07-17T09:17:08.3549892Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0045s] [ 53%] 2025-07-17T09:17:08.3550353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 53%] 2025-07-17T09:17:08.3550816Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0036s] [ 53%] 2025-07-17T09:17:08.3551271Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0040s] [ 53%] 2025-07-17T09:17:08.3551733Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0050s] [ 53%] 2025-07-17T09:17:08.3552194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0048s] [ 53%] 2025-07-17T09:17:08.3552649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0047s] [ 53%] 2025-07-17T09:17:08.3553116Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0046s] [ 53%] 2025-07-17T09:17:08.3553570Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0052s] [ 53%] 2025-07-17T09:17:08.3554100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0050s] [ 53%] 2025-07-17T09:17:08.3554762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3555484Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3556134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3556712Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3557306Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3557889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3558475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3559056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3559637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3560215Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3560789Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3561365Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3561996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3562682Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3563425Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3563999Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3564591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3565171Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3565755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3566336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3566918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3567513Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3568096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3568675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3569308Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3570000Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3570627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3571203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0010s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3571785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3572365Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3572940Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0018s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3573526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3574108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3574688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3575272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3575848Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3576506Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0014s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3577186Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3577821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3578403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3578985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3579562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3580134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3580716Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3581297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3581870Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3582453Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3583032Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_203_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 53%] 2025-07-17T09:17:08.3583499Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0051s] [ 53%] 2025-07-17T09:17:08.3584079Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0041s] [ 53%] 2025-07-17T09:17:08.3584593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0052s] [ 53%] 2025-07-17T09:17:08.3585104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0047s] [ 53%] 2025-07-17T09:17:08.3585572Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0039s] [ 53%] 2025-07-17T09:17:08.3586040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 53%] 2025-07-17T09:17:08.3586502Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0049s] [ 53%] 2025-07-17T09:17:08.3586961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0043s] [ 53%] 2025-07-17T09:17:08.3587423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0039s] [ 53%] 2025-07-17T09:17:08.3587877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0042s] [ 53%] 2025-07-17T09:17:08.3588332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0049s] [ 53%] 2025-07-17T09:17:08.3588782Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0044s] [ 53%] 2025-07-17T09:17:08.3589243Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0040s] [ 53%] 2025-07-17T09:17:08.3589698Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0039s] [ 53%] 2025-07-17T09:17:08.3590157Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0050s] [ 53%] 2025-07-17T09:17:08.3590610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0043s] [ 53%] 2025-07-17T09:17:08.3591181Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 53%] 2025-07-17T09:17:08.3591700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 53%] 2025-07-17T09:17:08.3592209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0052s] [ 53%] 2025-07-17T09:17:08.3592671Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0045s] [ 53%] 2025-07-17T09:17:08.3593138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0042s] [ 53%] 2025-07-17T09:17:08.3593605Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0042s] [ 53%] 2025-07-17T09:17:08.3594146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0051s] [ 53%] 2025-07-17T09:17:08.3594614Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0045s] [ 53%] 2025-07-17T09:17:08.3595075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0042s] [ 53%] 2025-07-17T09:17:08.3595531Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0044s] [ 53%] 2025-07-17T09:17:08.3595987Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0054s] [ 53%] 2025-07-17T09:17:08.3596441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0046s] [ 53%] 2025-07-17T09:17:08.3596914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0043s] [ 53%] 2025-07-17T09:17:08.3597375Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0050s] [ 53%] 2025-07-17T09:17:08.3597833Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0061s] [ 53%] 2025-07-17T09:17:08.3598507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0052s] [ 53%] 2025-07-17T09:17:08.3599028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0048s] [ 53%] 2025-07-17T09:17:08.3599494Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0047s] [ 53%] 2025-07-17T09:17:08.3599951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0059s] [ 53%] 2025-07-17T09:17:08.3600415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0052s] [ 53%] 2025-07-17T09:17:08.3600884Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0044s] [ 53%] 2025-07-17T09:17:08.3601344Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0044s] [ 53%] 2025-07-17T09:17:08.3601812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0055s] [ 53%] 2025-07-17T09:17:08.3602277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0049s] [ 53%] 2025-07-17T09:17:08.3602743Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0044s] [ 53%] 2025-07-17T09:17:08.3603202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0042s] [ 53%] 2025-07-17T09:17:08.3603664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0053s] [ 53%] 2025-07-17T09:17:08.3604122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0045s] [ 53%] 2025-07-17T09:17:08.3604593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0041s] [ 53%] 2025-07-17T09:17:08.3605054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0040s] [ 54%] 2025-07-17T09:17:08.3605668Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0050s] [ 54%] 2025-07-17T09:17:08.3606178Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0044s] [ 54%] 2025-07-17T09:17:08.3606763Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3607360Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3607941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3608520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3609107Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3609690Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3610270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3610855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3611433Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3612011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3612691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3613322Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3613952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3614534Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3615110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3615680Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3616268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3616852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3617434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3618016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3618604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3619185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3619875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3620509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3621159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3621744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3622322Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3622896Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3623475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3624061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3624642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3625214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3625801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3626380Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3627015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3627697Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3628330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0015s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3628921Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3629503Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3630087Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3630673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3631246Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3631824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3632398Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3632981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3633561Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3634269Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3634975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_256_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3635499Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0046s] [ 54%] 2025-07-17T09:17:08.3635967Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 54%] 2025-07-17T09:17:08.3636428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 54%] 2025-07-17T09:17:08.3636885Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 54%] 2025-07-17T09:17:08.3637349Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0034s] [ 54%] 2025-07-17T09:17:08.3637806Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 54%] 2025-07-17T09:17:08.3638272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 54%] 2025-07-17T09:17:08.3638728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 54%] 2025-07-17T09:17:08.3639183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 54%] 2025-07-17T09:17:08.3639635Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 54%] 2025-07-17T09:17:08.3642378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0037s] [ 54%] 2025-07-17T09:17:08.3643887Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 54%] 2025-07-17T09:17:08.3644354Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0034s] [ 54%] 2025-07-17T09:17:08.3644943Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0034s] [ 54%] 2025-07-17T09:17:08.3645442Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0034s] [ 54%] 2025-07-17T09:17:08.3645900Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0037s] [ 54%] 2025-07-17T09:17:08.3646368Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0038s] [ 54%] 2025-07-17T09:17:08.3646827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0037s] [ 54%] 2025-07-17T09:17:08.3647286Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 54%] 2025-07-17T09:17:08.3647747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0040s] [ 54%] 2025-07-17T09:17:08.3648213Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0037s] [ 54%] 2025-07-17T09:17:08.3648679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0035s] [ 54%] 2025-07-17T09:17:08.3649135Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0040s] [ 54%] 2025-07-17T09:17:08.3649597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0039s] [ 54%] 2025-07-17T09:17:08.3650060Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0037s] [ 54%] 2025-07-17T09:17:08.3650517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0044s] [ 54%] 2025-07-17T09:17:08.3651093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0047s] [ 54%] 2025-07-17T09:17:08.3651613Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 54%] 2025-07-17T09:17:08.3652197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 54%] 2025-07-17T09:17:08.3652654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 54%] 2025-07-17T09:17:08.3653123Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 54%] 2025-07-17T09:17:08.3653572Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0036s] [ 54%] 2025-07-17T09:17:08.3654039Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 54%] 2025-07-17T09:17:08.3654501Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0034s] [ 54%] 2025-07-17T09:17:08.3654956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 54%] 2025-07-17T09:17:08.3655412Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 54%] 2025-07-17T09:17:08.3655876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 PASSED [0.0033s] [ 54%] 2025-07-17T09:17:08.3656340Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 PASSED [0.0033s] [ 54%] 2025-07-17T09:17:08.3656798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 PASSED [0.0035s] [ 54%] 2025-07-17T09:17:08.3657262Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 PASSED [0.0036s] [ 54%] 2025-07-17T09:17:08.3657725Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 54%] 2025-07-17T09:17:08.3658243Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 54%] 2025-07-17T09:17:08.3658753Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 54%] 2025-07-17T09:17:08.3659318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 54%] 2025-07-17T09:17:08.3659785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 PASSED [0.0033s] [ 54%] 2025-07-17T09:17:08.3660242Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 PASSED [0.0033s] [ 54%] 2025-07-17T09:17:08.3660702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 PASSED [0.0035s] [ 54%] 2025-07-17T09:17:08.3661160Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_False_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 PASSED [0.0035s] [ 54%] 2025-07-17T09:17:08.3661755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3662345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3662926Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3663504Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3664087Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3664667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3665338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3665965Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3666667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3667244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3667812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3668385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3668964Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3669542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3670122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3670691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3671273Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 54%] 2025-07-17T09:17:08.3671851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3672487Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3673118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3673756Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3674489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3675067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0008s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3675644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3676220Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3676795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3677365Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3677936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3678514Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3679089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3679721Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3680385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3681169Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3681838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3682425Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3683002Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale0_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3683587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3684167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_False_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3684740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads0_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3685320Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_bfloat16_scale_l1_enable_gqa_True_n_heads1_cuda_bfloat16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3685895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3686470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3687037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3687665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale0_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3688343Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads0_cuda_float16 SKIPPED [0.0009s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3688967Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_False_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3689550Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads0_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3690122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_8_seq_len_q_4_seq_len_k_579_head_dim_8_is_causal_True_dropout_p_0_48_float16_scale_l1_enable_gqa_True_n_heads1_cuda_float16 SKIPPED [0.0007s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 55%] 2025-07-17T09:17:08.3690596Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [2.9559s] [ 55%] 2025-07-17T09:17:08.3691054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0012s] [ 55%] 2025-07-17T09:17:08.3691515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0665s] [ 55%] 2025-07-17T09:17:08.3691972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0010s] [ 55%] 2025-07-17T09:17:08.3692424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0814s] [ 55%] 2025-07-17T09:17:08.3692885Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0011s] [ 55%] 2025-07-17T09:17:08.3693344Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0676s] [ 55%] 2025-07-17T09:17:08.3693815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0010s] [ 55%] 2025-07-17T09:17:08.3694325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0519s] [ 55%] 2025-07-17T09:17:08.3694829Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3695286Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0423s] [ 55%] 2025-07-17T09:17:08.3695848Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3696311Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0530s] [ 55%] 2025-07-17T09:17:08.3696766Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3697225Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0445s] [ 55%] 2025-07-17T09:17:08.3697679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3698139Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0505s] [ 55%] 2025-07-17T09:17:08.3698596Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0010s] [ 55%] 2025-07-17T09:17:08.3699047Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0481s] [ 55%] 2025-07-17T09:17:08.3699504Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3699958Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0604s] [ 55%] 2025-07-17T09:17:08.3700414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0015s] [ 55%] 2025-07-17T09:17:08.3700871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0578s] [ 55%] 2025-07-17T09:17:08.3701328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0010s] [ 55%] 2025-07-17T09:17:08.3701846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0469s] [ 55%] 2025-07-17T09:17:08.3702344Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3702910Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0446s] [ 55%] 2025-07-17T09:17:08.3703365Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3703831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0478s] [ 55%] 2025-07-17T09:17:08.3704277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3704738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0473s] [ 55%] 2025-07-17T09:17:08.3705196Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3705648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0679s] [ 55%] 2025-07-17T09:17:08.3706106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0010s] [ 55%] 2025-07-17T09:17:08.3706559Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0646s] [ 55%] 2025-07-17T09:17:08.3707015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0010s] [ 55%] 2025-07-17T09:17:08.3707468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0685s] [ 55%] 2025-07-17T09:17:08.3707918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0010s] [ 55%] 2025-07-17T09:17:08.3708379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0671s] [ 55%] 2025-07-17T09:17:08.3708900Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0010s] [ 55%] 2025-07-17T09:17:08.3709404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0498s] [ 55%] 2025-07-17T09:17:08.3709903Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3710406Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0460s] [ 55%] 2025-07-17T09:17:08.3710860Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3711311Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0509s] [ 55%] 2025-07-17T09:17:08.3711780Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3712229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0486s] [ 55%] 2025-07-17T09:17:08.3712689Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3713136Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0534s] [ 55%] 2025-07-17T09:17:08.3713590Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0010s] [ 55%] 2025-07-17T09:17:08.3714100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0492s] [ 55%] 2025-07-17T09:17:08.3714551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3715005Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0535s] [ 55%] 2025-07-17T09:17:08.3715455Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0010s] [ 55%] 2025-07-17T09:17:08.3715976Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0504s] [ 55%] 2025-07-17T09:17:08.3716489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3716945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0487s] [ 55%] 2025-07-17T09:17:08.3717519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3717975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0441s] [ 55%] 2025-07-17T09:17:08.3718429Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3718880Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0474s] [ 55%] 2025-07-17T09:17:08.3719335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3719784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0456s] [ 55%] 2025-07-17T09:17:08.3720234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_32_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0009s] [ 55%] 2025-07-17T09:17:08.3720694Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0268s] [ 55%] 2025-07-17T09:17:08.3721146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 55%] 2025-07-17T09:17:08.3721601Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0260s] [ 55%] 2025-07-17T09:17:08.3722056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 55%] 2025-07-17T09:17:08.3722514Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0256s] [ 55%] 2025-07-17T09:17:08.3723012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 55%] 2025-07-17T09:17:08.3723524Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0254s] [ 55%] 2025-07-17T09:17:08.3723978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 55%] 2025-07-17T09:17:08.3724547Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0180s] [ 55%] 2025-07-17T09:17:08.3725002Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 55%] 2025-07-17T09:17:08.3725462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0174s] [ 55%] 2025-07-17T09:17:08.3725918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 55%] 2025-07-17T09:17:08.3726372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0186s] [ 55%] 2025-07-17T09:17:08.3726823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 55%] 2025-07-17T09:17:08.3727275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0184s] [ 55%] 2025-07-17T09:17:08.3727736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 55%] 2025-07-17T09:17:08.3728190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0192s] [ 55%] 2025-07-17T09:17:08.3728640Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 55%] 2025-07-17T09:17:08.3729103Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0221s] [ 55%] 2025-07-17T09:17:08.3729557Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 55%] 2025-07-17T09:17:08.3730017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0200s] [ 55%] 2025-07-17T09:17:08.3730521Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3731028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0239s] [ 56%] 2025-07-17T09:17:08.3731610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3732071Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0172s] [ 56%] 2025-07-17T09:17:08.3732523Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3732976Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0169s] [ 56%] 2025-07-17T09:17:08.3733437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3733888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0175s] [ 56%] 2025-07-17T09:17:08.3734346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3734795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0177s] [ 56%] 2025-07-17T09:17:08.3735253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_256_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3735715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0232s] [ 56%] 2025-07-17T09:17:08.3736164Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3736620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0224s] [ 56%] 2025-07-17T09:17:08.3737072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3737587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0233s] [ 56%] 2025-07-17T09:17:08.3738088Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3738597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0233s] [ 56%] 2025-07-17T09:17:08.3739104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3739554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0177s] [ 56%] 2025-07-17T09:17:08.3740010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3740463Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0170s] [ 56%] 2025-07-17T09:17:08.3740918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3741368Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0183s] [ 56%] 2025-07-17T09:17:08.3741824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3742282Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0178s] [ 56%] 2025-07-17T09:17:08.3742735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_256_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3743208Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0184s] [ 56%] 2025-07-17T09:17:08.3743658Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3744113Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0175s] [ 56%] 2025-07-17T09:17:08.3744621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3745120Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0185s] [ 56%] 2025-07-17T09:17:08.3745569Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3746136Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0185s] [ 56%] 2025-07-17T09:17:08.3746600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_64_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3747051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0171s] [ 56%] 2025-07-17T09:17:08.3747503Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3747955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0168s] [ 56%] 2025-07-17T09:17:08.3748415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_0_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3748862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_False_cuda_float16 PASSED [0.0171s] [ 56%] 2025-07-17T09:17:08.3749312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale0_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3749767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_False_cuda_float16 PASSED [0.0172s] [ 56%] 2025-07-17T09:17:08.3750216Z test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_nestedtensor_batch_size_8_max_seq_len_q_32_max_seq_len_kv_32_head_dim_8_dropout_p_0_1_float16_scale_l1_is_causal_True_cuda_float16 PASSED [0.0008s] [ 56%] 2025-07-17T09:17:08.3750556Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_different_dk_dv_cuda SKIPPED [0.0001s] (cuDNN Attention is not supported on this system) [ 56%] 2025-07-17T09:17:08.3751031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [1.5144s] [ 56%] 2025-07-17T09:17:08.3751502Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [1.0080s] [ 56%] 2025-07-17T09:17:08.3752048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9308s] [ 56%] 2025-07-17T09:17:08.3752574Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8813s] [ 56%] 2025-07-17T09:17:08.3753153Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.9237s] [ 56%] 2025-07-17T09:17:08.3753622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.9159s] [ 56%] 2025-07-17T09:17:08.3754330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8950s] [ 56%] 2025-07-17T09:17:08.3754799Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8929s] [ 56%] 2025-07-17T09:17:08.3755276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8992s] [ 56%] 2025-07-17T09:17:08.3755737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.9092s] [ 56%] 2025-07-17T09:17:08.3756209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8906s] [ 56%] 2025-07-17T09:17:08.3756673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9238s] [ 56%] 2025-07-17T09:17:08.3757137Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.9004s] [ 56%] 2025-07-17T09:17:08.3757603Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8908s] [ 56%] 2025-07-17T09:17:08.3758067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9329s] [ 56%] 2025-07-17T09:17:08.3758539Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8912s] [ 56%] 2025-07-17T09:17:08.3759003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8881s] [ 56%] 2025-07-17T09:17:08.3759539Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8833s] [ 56%] 2025-07-17T09:17:08.3760085Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9077s] [ 56%] 2025-07-17T09:17:08.3760702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8853s] [ 56%] 2025-07-17T09:17:08.3761175Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8880s] [ 56%] 2025-07-17T09:17:08.3761642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8894s] [ 56%] 2025-07-17T09:17:08.3762112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8908s] [ 56%] 2025-07-17T09:17:08.3762578Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8890s] [ 56%] 2025-07-17T09:17:08.3763049Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8960s] [ 56%] 2025-07-17T09:17:08.3763515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8825s] [ 56%] 2025-07-17T09:17:08.3763978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8868s] [ 56%] 2025-07-17T09:17:08.3764443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9235s] [ 56%] 2025-07-17T09:17:08.3764905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.9050s] [ 56%] 2025-07-17T09:17:08.3765370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8859s] [ 56%] 2025-07-17T09:17:08.3765830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8874s] [ 56%] 2025-07-17T09:17:08.3766300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9002s] [ 56%] 2025-07-17T09:17:08.3766825Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8871s] [ 56%] 2025-07-17T09:17:08.3767338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8802s] [ 56%] 2025-07-17T09:17:08.3767911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8887s] [ 56%] 2025-07-17T09:17:08.3768379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8872s] [ 56%] 2025-07-17T09:17:08.3768848Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8928s] [ 56%] 2025-07-17T09:17:08.3769308Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8929s] [ 56%] 2025-07-17T09:17:08.3769778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8864s] [ 56%] 2025-07-17T09:17:08.3770247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8854s] [ 56%] 2025-07-17T09:17:08.3770839Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 56%] 2025-07-17T09:17:08.3771305Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8922s] [ 56%] 2025-07-17T09:17:08.3771893Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 56%] 2025-07-17T09:17:08.3772360Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9207s] [ 56%] 2025-07-17T09:17:08.3772941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 56%] 2025-07-17T09:17:08.3773414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8847s] [ 56%] 2025-07-17T09:17:08.3774058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 56%] 2025-07-17T09:17:08.3774568Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8977s] [ 56%] 2025-07-17T09:17:08.3775139Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8924s] [ 56%] 2025-07-17T09:17:08.3775603Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8882s] [ 56%] 2025-07-17T09:17:08.3776075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9061s] [ 56%] 2025-07-17T09:17:08.3776534Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8837s] [ 56%] 2025-07-17T09:17:08.3777002Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8837s] [ 56%] 2025-07-17T09:17:08.3777469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8884s] [ 56%] 2025-07-17T09:17:08.3777935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9254s] [ 56%] 2025-07-17T09:17:08.3778403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8854s] [ 56%] 2025-07-17T09:17:08.3778999Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 56%] 2025-07-17T09:17:08.3779472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8852s] [ 56%] 2025-07-17T09:17:08.3780062Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 56%] 2025-07-17T09:17:08.3780522Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8876s] [ 56%] 2025-07-17T09:17:08.3781177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 56%] 2025-07-17T09:17:08.3781687Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8913s] [ 56%] 2025-07-17T09:17:08.3782325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0015s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 56%] 2025-07-17T09:17:08.3782841Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8934s] [ 56%] 2025-07-17T09:17:08.3783313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.9009s] [ 56%] 2025-07-17T09:17:08.3783778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8874s] [ 56%] 2025-07-17T09:17:08.3784239Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8884s] [ 56%] 2025-07-17T09:17:08.3784706Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8965s] [ 56%] 2025-07-17T09:17:08.3785169Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8994s] [ 56%] 2025-07-17T09:17:08.3785634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8889s] [ 56%] 2025-07-17T09:17:08.3786099Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8914s] [ 56%] 2025-07-17T09:17:08.3786566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8953s] [ 56%] 2025-07-17T09:17:08.3787157Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3787619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8922s] [ 57%] 2025-07-17T09:17:08.3788263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3788788Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8966s] [ 57%] 2025-07-17T09:17:08.3789432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3789953Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8934s] [ 57%] 2025-07-17T09:17:08.3790538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3791008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8905s] [ 57%] 2025-07-17T09:17:08.3791471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.9071s] [ 57%] 2025-07-17T09:17:08.3791937Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8929s] [ 57%] 2025-07-17T09:17:08.3792401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8888s] [ 57%] 2025-07-17T09:17:08.3792870Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8936s] [ 57%] 2025-07-17T09:17:08.3793338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.9030s] [ 57%] 2025-07-17T09:17:08.3793799Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8986s] [ 57%] 2025-07-17T09:17:08.3794332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8974s] [ 57%] 2025-07-17T09:17:08.3794797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9040s] [ 57%] 2025-07-17T09:17:08.3795389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3795951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.9043s] [ 57%] 2025-07-17T09:17:08.3796747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3797339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9180s] [ 57%] 2025-07-17T09:17:08.3797932Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3798403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.9086s] [ 57%] 2025-07-17T09:17:08.3798992Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3799462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8999s] [ 57%] 2025-07-17T09:17:08.3799927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.9008s] [ 57%] 2025-07-17T09:17:08.3800384Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.9268s] [ 57%] 2025-07-17T09:17:08.3800851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8942s] [ 57%] 2025-07-17T09:17:08.3801315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8915s] [ 57%] 2025-07-17T09:17:08.3801780Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8937s] [ 57%] 2025-07-17T09:17:08.3802244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8961s] [ 57%] 2025-07-17T09:17:08.3802715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9011s] [ 57%] 2025-07-17T09:17:08.3803230Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8989s] [ 57%] 2025-07-17T09:17:08.3803742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8828s] [ 57%] 2025-07-17T09:17:08.3804258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.9164s] [ 57%] 2025-07-17T09:17:08.3804765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9001s] [ 57%] 2025-07-17T09:17:08.3805232Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8849s] [ 57%] 2025-07-17T09:17:08.3805691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8899s] [ 57%] 2025-07-17T09:17:08.3806153Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8840s] [ 57%] 2025-07-17T09:17:08.3806619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8919s] [ 57%] 2025-07-17T09:17:08.3807082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8920s] [ 57%] 2025-07-17T09:17:08.3807549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8941s] [ 57%] 2025-07-17T09:17:08.3808010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8864s] [ 57%] 2025-07-17T09:17:08.3808472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9229s] [ 57%] 2025-07-17T09:17:08.3808932Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8826s] [ 57%] 2025-07-17T09:17:08.3809398Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8834s] [ 57%] 2025-07-17T09:17:08.3809857Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8857s] [ 57%] 2025-07-17T09:17:08.3810383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8901s] [ 57%] 2025-07-17T09:17:08.3810898Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8914s] [ 57%] 2025-07-17T09:17:08.3811409Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8915s] [ 57%] 2025-07-17T09:17:08.3811924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8857s] [ 57%] 2025-07-17T09:17:08.3812383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8905s] [ 57%] 2025-07-17T09:17:08.3812847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9012s] [ 57%] 2025-07-17T09:17:08.3813304Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8892s] [ 57%] 2025-07-17T09:17:08.3813767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8871s] [ 57%] 2025-07-17T09:17:08.3814235Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8888s] [ 57%] 2025-07-17T09:17:08.3814690Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_1_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8912s] [ 57%] 2025-07-17T09:17:08.3815161Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8938s] [ 57%] 2025-07-17T09:17:08.3815623Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8953s] [ 57%] 2025-07-17T09:17:08.3816091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9233s] [ 57%] 2025-07-17T09:17:08.3816556Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8916s] [ 57%] 2025-07-17T09:17:08.3817033Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8926s] [ 57%] 2025-07-17T09:17:08.3817581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8895s] [ 57%] 2025-07-17T09:17:08.3818101Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8986s] [ 57%] 2025-07-17T09:17:08.3818621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8988s] [ 57%] 2025-07-17T09:17:08.3819138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8933s] [ 57%] 2025-07-17T09:17:08.3819610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8922s] [ 57%] 2025-07-17T09:17:08.3820077Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9204s] [ 57%] 2025-07-17T09:17:08.3820551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8884s] [ 57%] 2025-07-17T09:17:08.3821017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8914s] [ 57%] 2025-07-17T09:17:08.3821483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8930s] [ 57%] 2025-07-17T09:17:08.3821952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8965s] [ 57%] 2025-07-17T09:17:08.3822419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9003s] [ 57%] 2025-07-17T09:17:08.3822894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8891s] [ 57%] 2025-07-17T09:17:08.3823356Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.9085s] [ 57%] 2025-07-17T09:17:08.3823821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8881s] [ 57%] 2025-07-17T09:17:08.3824293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8885s] [ 57%] 2025-07-17T09:17:08.3824816Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8894s] [ 57%] 2025-07-17T09:17:08.3825333Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8964s] [ 57%] 2025-07-17T09:17:08.3825854Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8924s] [ 57%] 2025-07-17T09:17:08.3826379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8877s] [ 57%] 2025-07-17T09:17:08.3826843Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8890s] [ 57%] 2025-07-17T09:17:08.3827312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8997s] [ 57%] 2025-07-17T09:17:08.3827782Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9056s] [ 57%] 2025-07-17T09:17:08.3828245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8885s] [ 57%] 2025-07-17T09:17:08.3828711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8901s] [ 57%] 2025-07-17T09:17:08.3829172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8918s] [ 57%] 2025-07-17T09:17:08.3829647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9067s] [ 57%] 2025-07-17T09:17:08.3830114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8931s] [ 57%] 2025-07-17T09:17:08.3830582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8851s] [ 57%] 2025-07-17T09:17:08.3831047Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8859s] [ 57%] 2025-07-17T09:17:08.3831511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9432s] [ 57%] 2025-07-17T09:17:08.3832036Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9423s] [ 57%] 2025-07-17T09:17:08.3832552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8898s] [ 57%] 2025-07-17T09:17:08.3833071Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8871s] [ 57%] 2025-07-17T09:17:08.3833591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8848s] [ 57%] 2025-07-17T09:17:08.3834131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8854s] [ 57%] 2025-07-17T09:17:08.3834729Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3835190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8873s] [ 57%] 2025-07-17T09:17:08.3835779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3836237Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8848s] [ 57%] 2025-07-17T09:17:08.3836831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3837297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8912s] [ 57%] 2025-07-17T09:17:08.3837879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3838347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9029s] [ 57%] 2025-07-17T09:17:08.3838809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.9297s] [ 57%] 2025-07-17T09:17:08.3839341Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8845s] [ 57%] 2025-07-17T09:17:08.3839862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8840s] [ 57%] 2025-07-17T09:17:08.3840392Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8823s] [ 57%] 2025-07-17T09:17:08.3840935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8885s] [ 57%] 2025-07-17T09:17:08.3841402Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8939s] [ 57%] 2025-07-17T09:17:08.3841872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8859s] [ 57%] 2025-07-17T09:17:08.3842336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8886s] [ 57%] 2025-07-17T09:17:08.3842919Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3843386Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.9310s] [ 57%] 2025-07-17T09:17:08.3843972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 57%] 2025-07-17T09:17:08.3844440Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9030s] [ 57%] 2025-07-17T09:17:08.3845031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 58%] 2025-07-17T09:17:08.3845494Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8962s] [ 58%] 2025-07-17T09:17:08.3846077Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0012s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 58%] 2025-07-17T09:17:08.3846607Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_1024_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8919s] [ 58%] 2025-07-17T09:17:08.3847124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8921s] [ 58%] 2025-07-17T09:17:08.3847637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8863s] [ 58%] 2025-07-17T09:17:08.3848151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8912s] [ 58%] 2025-07-17T09:17:08.3848615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8877s] [ 58%] 2025-07-17T09:17:08.3849082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8918s] [ 58%] 2025-07-17T09:17:08.3849547Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8907s] [ 58%] 2025-07-17T09:17:08.3850015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8830s] [ 58%] 2025-07-17T09:17:08.3850487Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8778s] [ 58%] 2025-07-17T09:17:08.3851069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 58%] 2025-07-17T09:17:08.3851532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8743s] [ 58%] 2025-07-17T09:17:08.3852121Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 58%] 2025-07-17T09:17:08.3852591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8740s] [ 58%] 2025-07-17T09:17:08.3853178Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 58%] 2025-07-17T09:17:08.3853702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8847s] [ 58%] 2025-07-17T09:17:08.3854342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 58%] 2025-07-17T09:17:08.3854857Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8779s] [ 58%] 2025-07-17T09:17:08.3855382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8803s] [ 58%] 2025-07-17T09:17:08.3855844Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.9178s] [ 58%] 2025-07-17T09:17:08.3856311Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8926s] [ 58%] 2025-07-17T09:17:08.3856778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8888s] [ 58%] 2025-07-17T09:17:08.3857247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.9457s] [ 58%] 2025-07-17T09:17:08.3857713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.9307s] [ 58%] 2025-07-17T09:17:08.3858178Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8970s] [ 58%] 2025-07-17T09:17:08.3858646Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8968s] [ 58%] 2025-07-17T09:17:08.3859226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 58%] 2025-07-17T09:17:08.3859732Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8974s] [ 58%] 2025-07-17T09:17:08.3860326Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 58%] 2025-07-17T09:17:08.3860855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8996s] [ 58%] 2025-07-17T09:17:08.3861496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 58%] 2025-07-17T09:17:08.3861957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8969s] [ 58%] 2025-07-17T09:17:08.3862653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 SKIPPED [0.0011s] (Flash V2 does not accept is_casual when seq_len_q != seq_len_k) [ 58%] 2025-07-17T09:17:08.3863115Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_1024_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8953s] [ 58%] 2025-07-17T09:17:08.3863583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8873s] [ 58%] 2025-07-17T09:17:08.3864049Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8949s] [ 58%] 2025-07-17T09:17:08.3864512Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9228s] [ 58%] 2025-07-17T09:17:08.3864976Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8950s] [ 58%] 2025-07-17T09:17:08.3865437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8873s] [ 58%] 2025-07-17T09:17:08.3865901Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.9027s] [ 58%] 2025-07-17T09:17:08.3866364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.9148s] [ 58%] 2025-07-17T09:17:08.3866831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8940s] [ 58%] 2025-07-17T09:17:08.3867291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8848s] [ 58%] 2025-07-17T09:17:08.3867749Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8814s] [ 58%] 2025-07-17T09:17:08.3868275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8871s] [ 58%] 2025-07-17T09:17:08.3868784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9094s] [ 58%] 2025-07-17T09:17:08.3869317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8840s] [ 58%] 2025-07-17T09:17:08.3869828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8821s] [ 58%] 2025-07-17T09:17:08.3870292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8887s] [ 58%] 2025-07-17T09:17:08.3870756Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_32_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8913s] [ 58%] 2025-07-17T09:17:08.3871219Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8894s] [ 58%] 2025-07-17T09:17:08.3871678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8895s] [ 58%] 2025-07-17T09:17:08.3872165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8930s] [ 58%] 2025-07-17T09:17:08.3872629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9135s] [ 58%] 2025-07-17T09:17:08.3873087Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8891s] [ 58%] 2025-07-17T09:17:08.3873554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8914s] [ 58%] 2025-07-17T09:17:08.3874084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8997s] [ 58%] 2025-07-17T09:17:08.3874546Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_False_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.9017s] [ 58%] 2025-07-17T09:17:08.3875010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8949s] [ 58%] 2025-07-17T09:17:08.3875533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8859s] [ 58%] 2025-07-17T09:17:08.3876063Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8780s] [ 58%] 2025-07-17T09:17:08.3876518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_0_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8859s] [ 58%] 2025-07-17T09:17:08.3877116Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel0_cuda_float16 PASSED [0.8831s] [ 58%] 2025-07-17T09:17:08.3877584Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale0_fused_kernel1_cuda_float16 PASSED [0.8815s] [ 58%] 2025-07-17T09:17:08.3878043Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel0_cuda_float16 PASSED [0.8835s] [ 58%] 2025-07-17T09:17:08.3878507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_attention_vs_math_ref_grads_cudagraph_batch_size_8_seq_len_q_256_seq_len_k_256_head_dim_64_is_causal_True_dropout_p_0_22_float16_scale_l1_fused_kernel1_cuda_float16 PASSED [0.8844s] [ 58%] 2025-07-17T09:17:08.3878954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_backwards_throws_determinism_warning_fused_kernel0_warn_only_False_cuda SKIPPED [0.0010s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 58%] 2025-07-17T09:17:08.3879392Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_backwards_throws_determinism_warning_fused_kernel0_warn_only_True_cuda SKIPPED [0.0007s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 58%] 2025-07-17T09:17:08.3879832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_backwards_throws_determinism_warning_fused_kernel1_warn_only_False_cuda SKIPPED [0.0007s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 58%] 2025-07-17T09:17:08.3880269Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_backwards_throws_determinism_warning_fused_kernel1_warn_only_True_cuda SKIPPED [0.0007s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 58%] 2025-07-17T09:17:08.3880803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0220s] [ 58%] 2025-07-17T09:17:08.3881336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0143s] [ 58%] 2025-07-17T09:17:08.3881863Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0142s] [ 58%] 2025-07-17T09:17:08.3882386Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0146s] [ 58%] 2025-07-17T09:17:08.3882962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0143s] [ 58%] 2025-07-17T09:17:08.3883525Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0146s] [ 58%] 2025-07-17T09:17:08.3884048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0146s] [ 58%] 2025-07-17T09:17:08.3884665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0104s] [ 58%] 2025-07-17T09:17:08.3885194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0184s] [ 58%] 2025-07-17T09:17:08.3885713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0138s] [ 58%] 2025-07-17T09:17:08.3886223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0139s] [ 58%] 2025-07-17T09:17:08.3886740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0141s] [ 58%] 2025-07-17T09:17:08.3887253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0139s] [ 58%] 2025-07-17T09:17:08.3887772Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0141s] [ 58%] 2025-07-17T09:17:08.3888294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0142s] [ 58%] 2025-07-17T09:17:08.3888806Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0097s] [ 58%] 2025-07-17T09:17:08.3889331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0135s] [ 58%] 2025-07-17T09:17:08.3889841Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0139s] [ 58%] 2025-07-17T09:17:08.3890373Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0138s] [ 58%] 2025-07-17T09:17:08.3890955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0141s] [ 58%] 2025-07-17T09:17:08.3891527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0138s] [ 58%] 2025-07-17T09:17:08.3892158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0143s] [ 58%] 2025-07-17T09:17:08.3892679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0143s] [ 58%] 2025-07-17T09:17:08.3893195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0097s] [ 58%] 2025-07-17T09:17:08.3893713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0216s] [ 58%] 2025-07-17T09:17:08.3894232Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0136s] [ 58%] 2025-07-17T09:17:08.3894754Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0133s] [ 58%] 2025-07-17T09:17:08.3895269Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0139s] [ 58%] 2025-07-17T09:17:08.3895786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0136s] [ 58%] 2025-07-17T09:17:08.3896300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0139s] [ 58%] 2025-07-17T09:17:08.3896817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0139s] [ 58%] 2025-07-17T09:17:08.3897326Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0095s] [ 58%] 2025-07-17T09:17:08.3897860Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0134s] [ 58%] 2025-07-17T09:17:08.3898432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0136s] [ 58%] 2025-07-17T09:17:08.3899022Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0136s] [ 58%] 2025-07-17T09:17:08.3899539Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0140s] [ 58%] 2025-07-17T09:17:08.3900157Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0136s] [ 58%] 2025-07-17T09:17:08.3900682Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0140s] [ 58%] 2025-07-17T09:17:08.3901194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0139s] [ 58%] 2025-07-17T09:17:08.3901712Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0100s] [ 58%] 2025-07-17T09:17:08.3902229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0140s] [ 58%] 2025-07-17T09:17:08.3902753Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0132s] [ 58%] 2025-07-17T09:17:08.3903266Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0132s] [ 58%] 2025-07-17T09:17:08.3903776Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0136s] [ 58%] 2025-07-17T09:17:08.3904296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0132s] [ 59%] 2025-07-17T09:17:08.3904808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0135s] [ 59%] 2025-07-17T09:17:08.3905324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0137s] [ 59%] 2025-07-17T09:17:08.3905836Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0095s] [ 59%] 2025-07-17T09:17:08.3906405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0129s] [ 59%] 2025-07-17T09:17:08.3906973Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0132s] [ 59%] 2025-07-17T09:17:08.3907592Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0132s] [ 59%] 2025-07-17T09:17:08.3908109Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0135s] [ 59%] 2025-07-17T09:17:08.3908627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0132s] [ 59%] 2025-07-17T09:17:08.3909141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0136s] [ 59%] 2025-07-17T09:17:08.3909659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0135s] [ 59%] 2025-07-17T09:17:08.3910165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0094s] [ 59%] 2025-07-17T09:17:08.3910693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0022s] [ 59%] 2025-07-17T09:17:08.3911204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0021s] [ 59%] 2025-07-17T09:17:08.3911717Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0021s] [ 59%] 2025-07-17T09:17:08.3912223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0021s] [ 59%] 2025-07-17T09:17:08.3912737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda PASSED [0.0021s] [ 59%] 2025-07-17T09:17:08.3913253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda PASSED [0.0021s] [ 59%] 2025-07-17T09:17:08.3914065Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda PASSED [0.0021s] [ 59%] 2025-07-17T09:17:08.3914648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel0_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda PASSED [0.0020s] [ 59%] 2025-07-17T09:17:08.3915344Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3916030Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3916659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3917279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3917904Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3918516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3919138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3919753Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3920370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3920992Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3921671Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3922338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0008s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3923015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3923678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3924292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3924904Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3925531Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3926149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3926763Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3927379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3927991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3928606Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3929300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0008s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3929963Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3930638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3931297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3931916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3933267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3934555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3935833Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3937151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3938429Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_False_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3939720Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3941014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0008s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3942307Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3943658Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3944996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3946385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3947673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3948956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3950244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3951546Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3952839Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3954233Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3955519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0008s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3956843Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3958129Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3959499Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_False_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3960874Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3962324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3963616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3964894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3966172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3967450Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3968727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3970001Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_False_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0008s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3971284Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3972563Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3973846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3975179Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_False_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3976507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3977925Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_False_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3979203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_False_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3980469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_kernel1_expand_q_batch_True_expand_k_batch_True_expand_v_batch_True_expand_q_num_heads_True_expand_k_num_heads_True_expand_v_num_heads_True_cuda SKIPPED [0.0007s] (head_dim != head_dim_v unsupported on ROCm for now) [ 59%] 2025-07-17T09:17:08.3981583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_nested_broadcasting_query_dense_cuda SKIPPED [0.0007s] (skipIfRocm: Efficient Attention on ROCM does not support head_dim != head_dim_v for now.) [ 59%] 2025-07-17T09:17:08.3982339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_seq_len_1_inputs_fused_kernel0_cuda PASSED [0.0193s] [ 59%] 2025-07-17T09:17:08.3982896Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_kernels_seq_len_1_inputs_fused_kernel1_cuda PASSED [0.0179s] [ 59%] 2025-07-17T09:17:08.3983415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_sdp_choice_type_dense_cuda PASSED [0.0012s] [ 59%] 2025-07-17T09:17:08.3983904Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_sdp_choice_type_nested_cuda PASSED [0.0011s] [ 59%] 2025-07-17T09:17:08.3984525Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_sdp_priority_order_use_compile_False_cuda SKIPPED [0.0001s] (cuDNN Attention is not supported on this system) [ 59%] 2025-07-17T09:17:08.3985274Z test_transformers.py::TestSDPACudaOnlyCUDA::test_fused_sdp_priority_order_use_compile_True_cuda SKIPPED [0.0001s] (cuDNN Attention is not supported on this system) [ 59%] 2025-07-17T09:17:08.3985951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_eff_attention_long_sequence_mask_float16_cuda_float16 PASSED [0.0147s] [ 59%] 2025-07-17T09:17:08.3986537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_eff_attention_long_sequence_mask_float32_cuda_float32 PASSED [0.0107s] [ 59%] 2025-07-17T09:17:08.3987094Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_eff_attention_non_contig_mask_bug_cuda PASSED [0.0176s] [ 59%] 2025-07-17T09:17:08.3987666Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_eff_attention_non_contiguous_mask_float16_cuda_float16 PASSED [0.0059s] [ 59%] 2025-07-17T09:17:08.3988252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_eff_attention_non_contiguous_mask_float32_cuda_float32 PASSED [0.0067s] [ 59%] 2025-07-17T09:17:08.3988962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_eff_backwards_determinism_cuda SKIPPED [0.0002s] (This test is not behaving deterministaclly non-deterministaclly on CI/CD) [ 59%] 2025-07-17T09:17:08.3989872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0198s] [ 59%] 2025-07-17T09:17:08.3990918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0052s] [ 59%] 2025-07-17T09:17:08.3991950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0170s] [ 59%] 2025-07-17T09:17:08.3993040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0052s] [ 59%] 2025-07-17T09:17:08.3994118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.3401s] [ 59%] 2025-07-17T09:17:08.3995089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0048s] [ 59%] 2025-07-17T09:17:08.3996063Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0167s] [ 59%] 2025-07-17T09:17:08.3997051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0048s] [ 59%] 2025-07-17T09:17:08.3998031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0153s] [ 59%] 2025-07-17T09:17:08.3999001Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0049s] [ 59%] 2025-07-17T09:17:08.3999980Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0123s] [ 59%] 2025-07-17T09:17:08.4000950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0049s] [ 59%] 2025-07-17T09:17:08.4001922Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0120s] [ 59%] 2025-07-17T09:17:08.4002895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 59%] 2025-07-17T09:17:08.4003862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0109s] [ 59%] 2025-07-17T09:17:08.4004825Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 59%] 2025-07-17T09:17:08.4005857Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0103s] [ 59%] 2025-07-17T09:17:08.4006889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 59%] 2025-07-17T09:17:08.4007926Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0160s] [ 59%] 2025-07-17T09:17:08.4008962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 60%] 2025-07-17T09:17:08.4009938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0148s] [ 60%] 2025-07-17T09:17:08.4010904Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 60%] 2025-07-17T09:17:08.4011866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0123s] [ 60%] 2025-07-17T09:17:08.4012827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0043s] [ 60%] 2025-07-17T09:17:08.4013799Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0086s] [ 60%] 2025-07-17T09:17:08.4014770Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 60%] 2025-07-17T09:17:08.4015739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0082s] [ 60%] 2025-07-17T09:17:08.4016715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 60%] 2025-07-17T09:17:08.4017676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0064s] [ 60%] 2025-07-17T09:17:08.4018635Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 60%] 2025-07-17T09:17:08.4019607Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0113s] [ 60%] 2025-07-17T09:17:08.4020659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0044s] [ 60%] 2025-07-17T09:17:08.4021682Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0105s] [ 60%] 2025-07-17T09:17:08.4022646Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 60%] 2025-07-17T09:17:08.4023718Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0071s] [ 60%] 2025-07-17T09:17:08.4024686Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 60%] 2025-07-17T09:17:08.4025652Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0177s] [ 60%] 2025-07-17T09:17:08.4026622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 60%] 2025-07-17T09:17:08.4027593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0159s] [ 60%] 2025-07-17T09:17:08.4028561Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 60%] 2025-07-17T09:17:08.4029528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0121s] [ 60%] 2025-07-17T09:17:08.4030491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 60%] 2025-07-17T09:17:08.4031458Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0178s] [ 60%] 2025-07-17T09:17:08.4032437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0044s] [ 60%] 2025-07-17T09:17:08.4033406Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0167s] [ 60%] 2025-07-17T09:17:08.4039830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 60%] 2025-07-17T09:17:08.4041015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0116s] [ 60%] 2025-07-17T09:17:08.4042062Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0045s] [ 60%] 2025-07-17T09:17:08.4043041Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0823s] [ 60%] 2025-07-17T09:17:08.4044196Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0068s] [ 60%] 2025-07-17T09:17:08.4045177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0071s] [ 60%] 2025-07-17T09:17:08.4046148Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0064s] [ 60%] 2025-07-17T09:17:08.4047105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0073s] [ 60%] 2025-07-17T09:17:08.4048074Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0076s] [ 60%] 2025-07-17T09:17:08.4049051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0092s] [ 60%] 2025-07-17T09:17:08.4050037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0080s] [ 60%] 2025-07-17T09:17:08.4051015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0092s] [ 60%] 2025-07-17T09:17:08.4051983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0078s] [ 60%] 2025-07-17T09:17:08.4052948Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0091s] [ 60%] 2025-07-17T09:17:08.4053911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0085s] [ 60%] 2025-07-17T09:17:08.4054888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 60%] 2025-07-17T09:17:08.4055855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 60%] 2025-07-17T09:17:08.4056891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 60%] 2025-07-17T09:17:08.4057911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 60%] 2025-07-17T09:17:08.4058996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0051s] [ 60%] 2025-07-17T09:17:08.4059973Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 60%] 2025-07-17T09:17:08.4060951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 60%] 2025-07-17T09:17:08.4061929Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0047s] [ 60%] 2025-07-17T09:17:08.4062903Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0048s] [ 60%] 2025-07-17T09:17:08.4063870Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0048s] [ 60%] 2025-07-17T09:17:08.4064830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0050s] [ 60%] 2025-07-17T09:17:08.4065786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0050s] [ 60%] 2025-07-17T09:17:08.4066754Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 60%] 2025-07-17T09:17:08.4067721Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 60%] 2025-07-17T09:17:08.4068688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0040s] [ 60%] 2025-07-17T09:17:08.4069646Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 60%] 2025-07-17T09:17:08.4070603Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0041s] [ 60%] 2025-07-17T09:17:08.4071617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 60%] 2025-07-17T09:17:08.4072636Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 60%] 2025-07-17T09:17:08.4073726Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 60%] 2025-07-17T09:17:08.4074799Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0047s] [ 60%] 2025-07-17T09:17:08.4075763Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0046s] [ 60%] 2025-07-17T09:17:08.4076725Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0049s] [ 60%] 2025-07-17T09:17:08.4077684Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0048s] [ 60%] 2025-07-17T09:17:08.4078651Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0561s] [ 60%] 2025-07-17T09:17:08.4079622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0076s] [ 60%] 2025-07-17T09:17:08.4080586Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0090s] [ 60%] 2025-07-17T09:17:08.4081548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0073s] [ 60%] 2025-07-17T09:17:08.4082504Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0093s] [ 60%] 2025-07-17T09:17:08.4083466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0076s] [ 60%] 2025-07-17T09:17:08.4084432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0111s] [ 60%] 2025-07-17T09:17:08.4085403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0089s] [ 60%] 2025-07-17T09:17:08.4086438Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0109s] [ 60%] 2025-07-17T09:17:08.4087461Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0072s] [ 60%] 2025-07-17T09:17:08.4088496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0104s] [ 60%] 2025-07-17T09:17:08.4089520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0086s] [ 60%] 2025-07-17T09:17:08.4090493Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0066s] [ 60%] 2025-07-17T09:17:08.4091470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0048s] [ 60%] 2025-07-17T09:17:08.4092438Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0062s] [ 60%] 2025-07-17T09:17:08.4093396Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0048s] [ 60%] 2025-07-17T09:17:08.4094363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0055s] [ 60%] 2025-07-17T09:17:08.4095325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0045s] [ 60%] 2025-07-17T09:17:08.4096295Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0062s] [ 60%] 2025-07-17T09:17:08.4097268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0053s] [ 60%] 2025-07-17T09:17:08.4098242Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0059s] [ 60%] 2025-07-17T09:17:08.4099202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0052s] [ 60%] 2025-07-17T09:17:08.4100169Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0050s] [ 60%] 2025-07-17T09:17:08.4101189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0049s] [ 60%] 2025-07-17T09:17:08.4102231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 60%] 2025-07-17T09:17:08.4103195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 60%] 2025-07-17T09:17:08.4104267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 60%] 2025-07-17T09:17:08.4105227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0044s] [ 60%] 2025-07-17T09:17:08.4106178Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0052s] [ 60%] 2025-07-17T09:17:08.4107140Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 60%] 2025-07-17T09:17:08.4108112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 60%] 2025-07-17T09:17:08.4109091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 60%] 2025-07-17T09:17:08.4110052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0043s] [ 60%] 2025-07-17T09:17:08.4111018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 60%] 2025-07-17T09:17:08.4111979Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0052s] [ 60%] 2025-07-17T09:17:08.4112945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 60%] 2025-07-17T09:17:08.4113989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 60%] 2025-07-17T09:17:08.4114958Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 60%] 2025-07-17T09:17:08.4115990Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 60%] 2025-07-17T09:17:08.4117010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 60%] 2025-07-17T09:17:08.4117958Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 60%] 2025-07-17T09:17:08.4119052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 60%] 2025-07-17T09:17:08.4120017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 60%] 2025-07-17T09:17:08.4120981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 60%] 2025-07-17T09:17:08.4121939Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 60%] 2025-07-17T09:17:08.4122898Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 60%] 2025-07-17T09:17:08.4123858Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0050s] [ 60%] 2025-07-17T09:17:08.4124806Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 60%] 2025-07-17T09:17:08.4125762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 60%] 2025-07-17T09:17:08.4126734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 60%] 2025-07-17T09:17:08.4127692Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0035s] [ 60%] 2025-07-17T09:17:08.4128663Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 61%] 2025-07-17T09:17:08.4129618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0045s] [ 61%] 2025-07-17T09:17:08.4130581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 61%] 2025-07-17T09:17:08.4131597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0050s] [ 61%] 2025-07-17T09:17:08.4132617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 61%] 2025-07-17T09:17:08.4133641Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0051s] [ 61%] 2025-07-17T09:17:08.4134724Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0043s] [ 61%] 2025-07-17T09:17:08.4135687Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0041s] [ 61%] 2025-07-17T09:17:08.4136648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 61%] 2025-07-17T09:17:08.4137622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0057s] [ 61%] 2025-07-17T09:17:08.4138587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 61%] 2025-07-17T09:17:08.4139548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0056s] [ 61%] 2025-07-17T09:17:08.4140502Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 61%] 2025-07-17T09:17:08.4141465Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0053s] [ 61%] 2025-07-17T09:17:08.4142415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 61%] 2025-07-17T09:17:08.4143389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0054s] [ 61%] 2025-07-17T09:17:08.4144369Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 61%] 2025-07-17T09:17:08.4145340Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0054s] [ 61%] 2025-07-17T09:17:08.4146348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 61%] 2025-07-17T09:17:08.4147349Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0051s] [ 61%] 2025-07-17T09:17:08.4148299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 61%] 2025-07-17T09:17:08.4149370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 61%] 2025-07-17T09:17:08.4150348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 61%] 2025-07-17T09:17:08.4151304Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 61%] 2025-07-17T09:17:08.4152254Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 61%] 2025-07-17T09:17:08.4153207Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 61%] 2025-07-17T09:17:08.4154250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 61%] 2025-07-17T09:17:08.4155227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 61%] 2025-07-17T09:17:08.4156192Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 61%] 2025-07-17T09:17:08.4157151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0051s] [ 61%] 2025-07-17T09:17:08.4158101Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 61%] 2025-07-17T09:17:08.4159059Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0042s] [ 61%] 2025-07-17T09:17:08.4160007Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 61%] 2025-07-17T09:17:08.4161041Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 61%] 2025-07-17T09:17:08.4162070Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 61%] 2025-07-17T09:17:08.4163025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 61%] 2025-07-17T09:17:08.4164123Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 61%] 2025-07-17T09:17:08.4165078Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 61%] 2025-07-17T09:17:08.4166029Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 61%] 2025-07-17T09:17:08.4166981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 61%] 2025-07-17T09:17:08.4167942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 61%] 2025-07-17T09:17:08.4168893Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0046s] [ 61%] 2025-07-17T09:17:08.4169842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 61%] 2025-07-17T09:17:08.4170784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 61%] 2025-07-17T09:17:08.4171730Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 61%] 2025-07-17T09:17:08.4172680Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0054s] [ 61%] 2025-07-17T09:17:08.4173636Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 61%] 2025-07-17T09:17:08.4174592Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0052s] [ 61%] 2025-07-17T09:17:08.4175551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 61%] 2025-07-17T09:17:08.4176576Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 61%] 2025-07-17T09:17:08.4177573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 61%] 2025-07-17T09:17:08.4178581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0059s] [ 61%] 2025-07-17T09:17:08.4179608Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 61%] 2025-07-17T09:17:08.4180569Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0058s] [ 61%] 2025-07-17T09:17:08.4181527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 61%] 2025-07-17T09:17:08.4182483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 61%] 2025-07-17T09:17:08.4183436Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 61%] 2025-07-17T09:17:08.4184401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0459s] [ 61%] 2025-07-17T09:17:08.4185377Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0059s] [ 61%] 2025-07-17T09:17:08.4186347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0058s] [ 61%] 2025-07-17T09:17:08.4187314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0058s] [ 61%] 2025-07-17T09:17:08.4188283Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0064s] [ 61%] 2025-07-17T09:17:08.4189242Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0064s] [ 61%] 2025-07-17T09:17:08.4190345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0043s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4191628Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4192896Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0023s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4194297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4195509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4196714Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4197803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0128s] [ 61%] 2025-07-17T09:17:08.4198783Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 61%] 2025-07-17T09:17:08.4199749Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0118s] [ 61%] 2025-07-17T09:17:08.4200710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 61%] 2025-07-17T09:17:08.4201677Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0102s] [ 61%] 2025-07-17T09:17:08.4202641Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 61%] 2025-07-17T09:17:08.4203724Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4204933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4206194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4207447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4208693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4209974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4211058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 61%] 2025-07-17T09:17:08.4212031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 61%] 2025-07-17T09:17:08.4213001Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0040s] [ 61%] 2025-07-17T09:17:08.4213952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 61%] 2025-07-17T09:17:08.4214906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 61%] 2025-07-17T09:17:08.4215863Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 61%] 2025-07-17T09:17:08.4216953Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4218160Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4219360Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4220548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4221801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4223064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0022s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4224245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0268s] [ 61%] 2025-07-17T09:17:08.4225226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0058s] [ 61%] 2025-07-17T09:17:08.4226195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0058s] [ 61%] 2025-07-17T09:17:08.4227153Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0057s] [ 61%] 2025-07-17T09:17:08.4228112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0060s] [ 61%] 2025-07-17T09:17:08.4229072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0061s] [ 61%] 2025-07-17T09:17:08.4230155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0032s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4231361Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4232569Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4233760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0043s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4235035Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4236297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4237444Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0055s] [ 61%] 2025-07-17T09:17:08.4238426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0068s] [ 61%] 2025-07-17T09:17:08.4239530Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0067s] [ 61%] 2025-07-17T09:17:08.4240500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0067s] [ 61%] 2025-07-17T09:17:08.4241471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0091s] [ 61%] 2025-07-17T09:17:08.4242431Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0088s] [ 61%] 2025-07-17T09:17:08.4243522Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0039s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4244728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0033s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4245927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4247131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4248337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 61%] 2025-07-17T09:17:08.4249533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0048s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4250614Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0054s] [ 62%] 2025-07-17T09:17:08.4251637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0044s] [ 62%] 2025-07-17T09:17:08.4252650Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0050s] [ 62%] 2025-07-17T09:17:08.4253608Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0043s] [ 62%] 2025-07-17T09:17:08.4254674Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0061s] [ 62%] 2025-07-17T09:17:08.4255647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0051s] [ 62%] 2025-07-17T09:17:08.4256732Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4257941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4259145Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4260337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4261530Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4262722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4263800Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 62%] 2025-07-17T09:17:08.4264774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0044s] [ 62%] 2025-07-17T09:17:08.4265740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0044s] [ 62%] 2025-07-17T09:17:08.4266755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0044s] [ 62%] 2025-07-17T09:17:08.4267763Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0049s] [ 62%] 2025-07-17T09:17:08.4268792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0049s] [ 62%] 2025-07-17T09:17:08.4269938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4271152Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4272354Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4273555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4274824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4276015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4277105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0312s] [ 62%] 2025-07-17T09:17:08.4278075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0067s] [ 62%] 2025-07-17T09:17:08.4279045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0071s] [ 62%] 2025-07-17T09:17:08.4280004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0066s] [ 62%] 2025-07-17T09:17:08.4280963Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0088s] [ 62%] 2025-07-17T09:17:08.4281979Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0088s] [ 62%] 2025-07-17T09:17:08.4283136Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0043s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4284468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4285673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4286866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0038s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4288070Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4289262Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4290350Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0051s] [ 62%] 2025-07-17T09:17:08.4291323Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 62%] 2025-07-17T09:17:08.4292291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0047s] [ 62%] 2025-07-17T09:17:08.4293253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 62%] 2025-07-17T09:17:08.4294215Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0049s] [ 62%] 2025-07-17T09:17:08.4295178Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 62%] 2025-07-17T09:17:08.4296318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4297578Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4298838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4300091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4301296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4302498Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4303578Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0051s] [ 62%] 2025-07-17T09:17:08.4304551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 62%] 2025-07-17T09:17:08.4305512Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0047s] [ 62%] 2025-07-17T09:17:08.4306467Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 62%] 2025-07-17T09:17:08.4307428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0055s] [ 62%] 2025-07-17T09:17:08.4308382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 62%] 2025-07-17T09:17:08.4309460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4310667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4311927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4313177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4314549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4315752Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4316831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 62%] 2025-07-17T09:17:08.4317797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 62%] 2025-07-17T09:17:08.4318767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 62%] 2025-07-17T09:17:08.4319724Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 62%] 2025-07-17T09:17:08.4320674Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 62%] 2025-07-17T09:17:08.4321621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 62%] 2025-07-17T09:17:08.4322698Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4323900Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4325102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4326363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4327628Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4328870Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4330000Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 62%] 2025-07-17T09:17:08.4330973Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 62%] 2025-07-17T09:17:08.4331935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0040s] [ 62%] 2025-07-17T09:17:08.4332897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 62%] 2025-07-17T09:17:08.4333850Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0041s] [ 62%] 2025-07-17T09:17:08.4334802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 62%] 2025-07-17T09:17:08.4335883Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4337082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4338286Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4339475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4340662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4341905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4343033Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 62%] 2025-07-17T09:17:08.4344130Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 62%] 2025-07-17T09:17:08.4345094Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 62%] 2025-07-17T09:17:08.4346056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 62%] 2025-07-17T09:17:08.4347012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0046s] [ 62%] 2025-07-17T09:17:08.4347966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 62%] 2025-07-17T09:17:08.4349048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4350248Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4351443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4352633Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4353823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4355080Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4356154Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0050s] [ 62%] 2025-07-17T09:17:08.4357211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 62%] 2025-07-17T09:17:08.4358231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0050s] [ 62%] 2025-07-17T09:17:08.4359305Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 62%] 2025-07-17T09:17:08.4360253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 62%] 2025-07-17T09:17:08.4361195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 62%] 2025-07-17T09:17:08.4362270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4363478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4364676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4365859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4367033Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4368209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4369275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 62%] 2025-07-17T09:17:08.4370227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 62%] 2025-07-17T09:17:08.4371180Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 62%] 2025-07-17T09:17:08.4372182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 62%] 2025-07-17T09:17:08.4373183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 62%] 2025-07-17T09:17:08.4374417Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 62%] 2025-07-17T09:17:08.4375505Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 62%] 2025-07-17T09:17:08.4376696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 63%] 2025-07-17T09:17:08.4377893Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 63%] 2025-07-17T09:17:08.4379071Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 63%] 2025-07-17T09:17:08.4380259Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 63%] 2025-07-17T09:17:08.4381437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 63%] 2025-07-17T09:17:08.4382503Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 63%] 2025-07-17T09:17:08.4383464Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 63%] 2025-07-17T09:17:08.4384423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 63%] 2025-07-17T09:17:08.4385383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 63%] 2025-07-17T09:17:08.4386335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 63%] 2025-07-17T09:17:08.4387356Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 63%] 2025-07-17T09:17:08.4388484Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 63%] 2025-07-17T09:17:08.4389817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 63%] 2025-07-17T09:17:08.4391016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 63%] 2025-07-17T09:17:08.4392203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 63%] 2025-07-17T09:17:08.4393398Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 63%] 2025-07-17T09:17:08.4394802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 63%] 2025-07-17T09:17:08.4395886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 63%] 2025-07-17T09:17:08.4396859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 63%] 2025-07-17T09:17:08.4397843Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 63%] 2025-07-17T09:17:08.4398812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 63%] 2025-07-17T09:17:08.4399774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 63%] 2025-07-17T09:17:08.4400737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 63%] 2025-07-17T09:17:08.4401704Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0058s] [ 63%] 2025-07-17T09:17:08.4402763Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 63%] 2025-07-17T09:17:08.4403804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0051s] [ 63%] 2025-07-17T09:17:08.4404907Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 63%] 2025-07-17T09:17:08.4405883Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 63%] 2025-07-17T09:17:08.4406846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 63%] 2025-07-17T09:17:08.4407808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 63%] 2025-07-17T09:17:08.4408777Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 63%] 2025-07-17T09:17:08.4409738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 63%] 2025-07-17T09:17:08.4410695Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 63%] 2025-07-17T09:17:08.4411649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0045s] [ 63%] 2025-07-17T09:17:08.4412612Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 63%] 2025-07-17T09:17:08.4413580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0048s] [ 63%] 2025-07-17T09:17:08.4414553Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 63%] 2025-07-17T09:17:08.4415519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0047s] [ 63%] 2025-07-17T09:17:08.4416477Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 63%] 2025-07-17T09:17:08.4417518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0039s] [ 63%] 2025-07-17T09:17:08.4418528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 63%] 2025-07-17T09:17:08.4419603Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 63%] 2025-07-17T09:17:08.4420566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 63%] 2025-07-17T09:17:08.4421526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 63%] 2025-07-17T09:17:08.4422478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 63%] 2025-07-17T09:17:08.4423444Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0046s] [ 63%] 2025-07-17T09:17:08.4424400Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 63%] 2025-07-17T09:17:08.4425362Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 63%] 2025-07-17T09:17:08.4426329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 63%] 2025-07-17T09:17:08.4427294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0046s] [ 63%] 2025-07-17T09:17:08.4428255Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 63%] 2025-07-17T09:17:08.4429210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 63%] 2025-07-17T09:17:08.4430159Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 63%] 2025-07-17T09:17:08.4431129Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0087s] [ 63%] 2025-07-17T09:17:08.4432152Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 63%] 2025-07-17T09:17:08.4433172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0075s] [ 63%] 2025-07-17T09:17:08.4434206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 63%] 2025-07-17T09:17:08.4435312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0073s] [ 63%] 2025-07-17T09:17:08.4436277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 63%] 2025-07-17T09:17:08.4437241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0102s] [ 63%] 2025-07-17T09:17:08.4438210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 63%] 2025-07-17T09:17:08.4439173Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0096s] [ 63%] 2025-07-17T09:17:08.4440135Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 63%] 2025-07-17T09:17:08.4441093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0092s] [ 63%] 2025-07-17T09:17:08.4442061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 63%] 2025-07-17T09:17:08.4443028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 63%] 2025-07-17T09:17:08.4444000Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 63%] 2025-07-17T09:17:08.4444972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 63%] 2025-07-17T09:17:08.4445939Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 63%] 2025-07-17T09:17:08.4446964Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 63%] 2025-07-17T09:17:08.4447991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 63%] 2025-07-17T09:17:08.4448959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0054s] [ 63%] 2025-07-17T09:17:08.4450051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 63%] 2025-07-17T09:17:08.4451026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0051s] [ 63%] 2025-07-17T09:17:08.4451990Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 63%] 2025-07-17T09:17:08.4452956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0043s] [ 63%] 2025-07-17T09:17:08.4453926Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0043s] [ 63%] 2025-07-17T09:17:08.4454895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 63%] 2025-07-17T09:17:08.4455862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 63%] 2025-07-17T09:17:08.4456823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 63%] 2025-07-17T09:17:08.4457785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 63%] 2025-07-17T09:17:08.4458744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 63%] 2025-07-17T09:17:08.4459697Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 63%] 2025-07-17T09:17:08.4460663Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0048s] [ 63%] 2025-07-17T09:17:08.4461126Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 63%] 2025-07-17T09:17:08.4461635Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0046s] [ 63%] 2025-07-17T09:17:08.4462137Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 63%] 2025-07-17T09:17:08.4462694Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 63%] 2025-07-17T09:17:08.4463150Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 63%] 2025-07-17T09:17:08.4463609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 63%] 2025-07-17T09:17:08.4464058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 63%] 2025-07-17T09:17:08.4464512Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 63%] 2025-07-17T09:17:08.4464961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 63%] 2025-07-17T09:17:08.4465415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 63%] 2025-07-17T09:17:08.4465866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 63%] 2025-07-17T09:17:08.4466318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 63%] 2025-07-17T09:17:08.4466781Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 63%] 2025-07-17T09:17:08.4467223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0038s] [ 63%] 2025-07-17T09:17:08.4467680Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 63%] 2025-07-17T09:17:08.4468128Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0039s] [ 63%] 2025-07-17T09:17:08.4468667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 63%] 2025-07-17T09:17:08.4469185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 63%] 2025-07-17T09:17:08.4469696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 63%] 2025-07-17T09:17:08.4470202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 63%] 2025-07-17T09:17:08.4470654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 63%] 2025-07-17T09:17:08.4471107Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 63%] 2025-07-17T09:17:08.4471556Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 63%] 2025-07-17T09:17:08.4472011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 63%] 2025-07-17T09:17:08.4472476Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 63%] 2025-07-17T09:17:08.4472925Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0042s] [ 63%] 2025-07-17T09:17:08.4473383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 63%] 2025-07-17T09:17:08.4473832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0044s] [ 63%] 2025-07-17T09:17:08.4474362Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0044s] [ 63%] 2025-07-17T09:17:08.4474809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0060s] [ 63%] 2025-07-17T09:17:08.4475274Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 63%] 2025-07-17T09:17:08.4475788Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0053s] [ 63%] 2025-07-17T09:17:08.4476301Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0043s] [ 64%] 2025-07-17T09:17:08.4476756Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0049s] [ 64%] 2025-07-17T09:17:08.4477346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 64%] 2025-07-17T09:17:08.4477808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0050s] [ 64%] 2025-07-17T09:17:08.4478267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0049s] [ 64%] 2025-07-17T09:17:08.4478717Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0048s] [ 64%] 2025-07-17T09:17:08.4479172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0048s] [ 64%] 2025-07-17T09:17:08.4479619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0047s] [ 64%] 2025-07-17T09:17:08.4480073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 64%] 2025-07-17T09:17:08.4480525Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 64%] 2025-07-17T09:17:08.4480982Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4481428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0034s] [ 64%] 2025-07-17T09:17:08.4481879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 64%] 2025-07-17T09:17:08.4482334Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 64%] 2025-07-17T09:17:08.4482836Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 64%] 2025-07-17T09:17:08.4483341Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0050s] [ 64%] 2025-07-17T09:17:08.4483793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 64%] 2025-07-17T09:17:08.4484351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0048s] [ 64%] 2025-07-17T09:17:08.4484804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 64%] 2025-07-17T09:17:08.4485253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0038s] [ 64%] 2025-07-17T09:17:08.4485696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 64%] 2025-07-17T09:17:08.4486154Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 64%] 2025-07-17T09:17:08.4486607Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4487054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4487502Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4487945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4488394Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 64%] 2025-07-17T09:17:08.4488839Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 64%] 2025-07-17T09:17:08.4489296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 64%] 2025-07-17T09:17:08.4489742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0045s] [ 64%] 2025-07-17T09:17:08.4490247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 64%] 2025-07-17T09:17:08.4490743Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 64%] 2025-07-17T09:17:08.4491240Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 64%] 2025-07-17T09:17:08.4491744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 64%] 2025-07-17T09:17:08.4492194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 64%] 2025-07-17T09:17:08.4492646Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0045s] [ 64%] 2025-07-17T09:17:08.4493122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 64%] 2025-07-17T09:17:08.4493572Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 64%] 2025-07-17T09:17:08.4494026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4494475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 64%] 2025-07-17T09:17:08.4494935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 64%] 2025-07-17T09:17:08.4495386Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0051s] [ 64%] 2025-07-17T09:17:08.4495842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 64%] 2025-07-17T09:17:08.4496280Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0042s] [ 64%] 2025-07-17T09:17:08.4496733Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 64%] 2025-07-17T09:17:08.4497239Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 64%] 2025-07-17T09:17:08.4497736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4498183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 64%] 2025-07-17T09:17:08.4498739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4499191Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4499633Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 64%] 2025-07-17T09:17:08.4500083Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 64%] 2025-07-17T09:17:08.4500543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 64%] 2025-07-17T09:17:08.4500991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0052s] [ 64%] 2025-07-17T09:17:08.4501445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 64%] 2025-07-17T09:17:08.4501890Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0042s] [ 64%] 2025-07-17T09:17:08.4502340Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 64%] 2025-07-17T09:17:08.4502782Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0048s] [ 64%] 2025-07-17T09:17:08.4503239Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4503689Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0046s] [ 64%] 2025-07-17T09:17:08.4504133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 64%] 2025-07-17T09:17:08.4504648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0039s] [ 64%] 2025-07-17T09:17:08.4505157Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 64%] 2025-07-17T09:17:08.4505717Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0051s] [ 64%] 2025-07-17T09:17:08.4506173Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 64%] 2025-07-17T09:17:08.4506617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0053s] [ 64%] 2025-07-17T09:17:08.4507067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 64%] 2025-07-17T09:17:08.4507511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 64%] 2025-07-17T09:17:08.4507957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 64%] 2025-07-17T09:17:08.4508401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 64%] 2025-07-17T09:17:08.4508849Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 64%] 2025-07-17T09:17:08.4509292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 64%] 2025-07-17T09:17:08.4509744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 64%] 2025-07-17T09:17:08.4510187Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4510635Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 64%] 2025-07-17T09:17:08.4511087Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 64%] 2025-07-17T09:17:08.4511587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 64%] 2025-07-17T09:17:08.4512082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0045s] [ 64%] 2025-07-17T09:17:08.4512522Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 64%] 2025-07-17T09:17:08.4513075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 64%] 2025-07-17T09:17:08.4513530Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 64%] 2025-07-17T09:17:08.4514058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0050s] [ 64%] 2025-07-17T09:17:08.4514520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 64%] 2025-07-17T09:17:08.4514964Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0049s] [ 64%] 2025-07-17T09:17:08.4515412Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 64%] 2025-07-17T09:17:08.4515848Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0043s] [ 64%] 2025-07-17T09:17:08.4516298Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 64%] 2025-07-17T09:17:08.4516741Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 64%] 2025-07-17T09:17:08.4517196Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 64%] 2025-07-17T09:17:08.4517639Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0043s] [ 64%] 2025-07-17T09:17:08.4518084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 64%] 2025-07-17T09:17:08.4518526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0046s] [ 64%] 2025-07-17T09:17:08.4519038Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 64%] 2025-07-17T09:17:08.4519562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0079s] [ 64%] 2025-07-17T09:17:08.4520138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4520592Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0073s] [ 64%] 2025-07-17T09:17:08.4521047Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4521491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0072s] [ 64%] 2025-07-17T09:17:08.4521943Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 64%] 2025-07-17T09:17:08.4522390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0100s] [ 64%] 2025-07-17T09:17:08.4522855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 64%] 2025-07-17T09:17:08.4523299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0095s] [ 64%] 2025-07-17T09:17:08.4523760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 64%] 2025-07-17T09:17:08.4524205Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0080s] [ 64%] 2025-07-17T09:17:08.4524656Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 64%] 2025-07-17T09:17:08.4525108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 64%] 2025-07-17T09:17:08.4525560Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 64%] 2025-07-17T09:17:08.4526061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 64%] 2025-07-17T09:17:08.4526558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 64%] 2025-07-17T09:17:08.4527059Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 64%] 2025-07-17T09:17:08.4527552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 64%] 2025-07-17T09:17:08.4528005Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 64%] 2025-07-17T09:17:08.4528461Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 64%] 2025-07-17T09:17:08.4528907Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0043s] [ 64%] 2025-07-17T09:17:08.4529356Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 64%] 2025-07-17T09:17:08.4529796Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0051s] [ 64%] 2025-07-17T09:17:08.4530252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 65%] 2025-07-17T09:17:08.4530697Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 65%] 2025-07-17T09:17:08.4531151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 65%] 2025-07-17T09:17:08.4531593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 65%] 2025-07-17T09:17:08.4532038Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 65%] 2025-07-17T09:17:08.4532481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4532995Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 65%] 2025-07-17T09:17:08.4533549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4534023Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4534583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4535041Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 65%] 2025-07-17T09:17:08.4535487Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0046s] [ 65%] 2025-07-17T09:17:08.4535935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 65%] 2025-07-17T09:17:08.4536385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 65%] 2025-07-17T09:17:08.4536844Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4537282Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4537734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 65%] 2025-07-17T09:17:08.4538188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0039s] [ 65%] 2025-07-17T09:17:08.4538643Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 65%] 2025-07-17T09:17:08.4539099Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 65%] 2025-07-17T09:17:08.4539552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4540002Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0044s] [ 65%] 2025-07-17T09:17:08.4540514Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4541013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0051s] [ 65%] 2025-07-17T09:17:08.4541564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 65%] 2025-07-17T09:17:08.4542025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 65%] 2025-07-17T09:17:08.4542487Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4542931Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4543388Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4543828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 65%] 2025-07-17T09:17:08.4544290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 65%] 2025-07-17T09:17:08.4544737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4545205Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 65%] 2025-07-17T09:17:08.4545660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4546114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4546562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4547011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4547521Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 65%] 2025-07-17T09:17:08.4548017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 65%] 2025-07-17T09:17:08.4548469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0034s] [ 65%] 2025-07-17T09:17:08.4549026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4549472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0044s] [ 65%] 2025-07-17T09:17:08.4549934Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 65%] 2025-07-17T09:17:08.4550383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 65%] 2025-07-17T09:17:08.4550841Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4551285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0045s] [ 65%] 2025-07-17T09:17:08.4551747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 65%] 2025-07-17T09:17:08.4552197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0044s] [ 65%] 2025-07-17T09:17:08.4552650Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 65%] 2025-07-17T09:17:08.4553102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 65%] 2025-07-17T09:17:08.4553555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4554073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4554587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4555097Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 65%] 2025-07-17T09:17:08.4555544Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 65%] 2025-07-17T09:17:08.4556143Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 65%] 2025-07-17T09:17:08.4556607Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4557052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 65%] 2025-07-17T09:17:08.4557507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4557952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4558407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4558855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4559308Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 65%] 2025-07-17T09:17:08.4559756Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0034s] [ 65%] 2025-07-17T09:17:08.4560202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 65%] 2025-07-17T09:17:08.4560648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 65%] 2025-07-17T09:17:08.4561096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 65%] 2025-07-17T09:17:08.4561551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 65%] 2025-07-17T09:17:08.4562057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 65%] 2025-07-17T09:17:08.4562555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 65%] 2025-07-17T09:17:08.4563057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 65%] 2025-07-17T09:17:08.4563544Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4563994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 65%] 2025-07-17T09:17:08.4564442Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 65%] 2025-07-17T09:17:08.4564899Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4565342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4565790Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4566231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4566684Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 65%] 2025-07-17T09:17:08.4567137Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4567593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4568038Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4568497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4569002Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0047s] [ 65%] 2025-07-17T09:17:08.4569493Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 65%] 2025-07-17T09:17:08.4569940Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 65%] 2025-07-17T09:17:08.4570500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 65%] 2025-07-17T09:17:08.4570947Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4571398Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4571835Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0032s] [ 65%] 2025-07-17T09:17:08.4572287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 65%] 2025-07-17T09:17:08.4572737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4573196Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4573644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4574089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4574535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 65%] 2025-07-17T09:17:08.4574976Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 65%] 2025-07-17T09:17:08.4575424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 65%] 2025-07-17T09:17:08.4575869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4576378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4576868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 65%] 2025-07-17T09:17:08.4577416Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4577873Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 65%] 2025-07-17T09:17:08.4578318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4578776Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4579220Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4579669Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 65%] 2025-07-17T09:17:08.4580106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4580556Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 65%] 2025-07-17T09:17:08.4581007Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 65%] 2025-07-17T09:17:08.4581456Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4581900Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4582340Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 65%] 2025-07-17T09:17:08.4582785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0049s] [ 65%] 2025-07-17T09:17:08.4583283Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 65%] 2025-07-17T09:17:08.4583786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 65%] 2025-07-17T09:17:08.4584244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 66%] 2025-07-17T09:17:08.4584813Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 66%] 2025-07-17T09:17:08.4585270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 66%] 2025-07-17T09:17:08.4585711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 66%] 2025-07-17T09:17:08.4586160Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 66%] 2025-07-17T09:17:08.4586609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 66%] 2025-07-17T09:17:08.4587059Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 66%] 2025-07-17T09:17:08.4587505Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0045s] [ 66%] 2025-07-17T09:17:08.4587949Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 66%] 2025-07-17T09:17:08.4588390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0042s] [ 66%] 2025-07-17T09:17:08.4588834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 66%] 2025-07-17T09:17:08.4589285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 66%] 2025-07-17T09:17:08.4589734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 66%] 2025-07-17T09:17:08.4590182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0049s] [ 66%] 2025-07-17T09:17:08.4590702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 66%] 2025-07-17T09:17:08.4591195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0043s] [ 66%] 2025-07-17T09:17:08.4591744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 66%] 2025-07-17T09:17:08.4592190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 66%] 2025-07-17T09:17:08.4592640Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 66%] 2025-07-17T09:17:08.4593078Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0045s] [ 66%] 2025-07-17T09:17:08.4593524Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 66%] 2025-07-17T09:17:08.4594075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0046s] [ 66%] 2025-07-17T09:17:08.4594523Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 66%] 2025-07-17T09:17:08.4594975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 66%] 2025-07-17T09:17:08.4595425Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 66%] 2025-07-17T09:17:08.4595869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0042s] [ 66%] 2025-07-17T09:17:08.4596308Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 66%] 2025-07-17T09:17:08.4596748Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0041s] [ 66%] 2025-07-17T09:17:08.4597189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 66%] 2025-07-17T09:17:08.4597703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 66%] 2025-07-17T09:17:08.4598208Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 66%] 2025-07-17T09:17:08.4598642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 66%] 2025-07-17T09:17:08.4599212Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 66%] 2025-07-17T09:17:08.4599654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0041s] [ 66%] 2025-07-17T09:17:08.4600110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 66%] 2025-07-17T09:17:08.4600552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 66%] 2025-07-17T09:17:08.4601008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 66%] 2025-07-17T09:17:08.4601444Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 66%] 2025-07-17T09:17:08.4601891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 66%] 2025-07-17T09:17:08.4602328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 66%] 2025-07-17T09:17:08.4602768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 66%] 2025-07-17T09:17:08.4603215Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 66%] 2025-07-17T09:17:08.4603659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 66%] 2025-07-17T09:17:08.4604101Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 66%] 2025-07-17T09:17:08.4604542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 66%] 2025-07-17T09:17:08.4605036Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0046s] [ 66%] 2025-07-17T09:17:08.4605532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 66%] 2025-07-17T09:17:08.4606023Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0052s] [ 66%] 2025-07-17T09:17:08.4606524Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 66%] 2025-07-17T09:17:08.4606962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0050s] [ 66%] 2025-07-17T09:17:08.4607415Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 66%] 2025-07-17T09:17:08.4607852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0051s] [ 66%] 2025-07-17T09:17:08.4608299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 66%] 2025-07-17T09:17:08.4608769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0667s] [ 66%] 2025-07-17T09:17:08.4609232Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0095s] [ 66%] 2025-07-17T09:17:08.4609691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0114s] [ 66%] 2025-07-17T09:17:08.4610145Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0093s] [ 66%] 2025-07-17T09:17:08.4610598Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0125s] [ 66%] 2025-07-17T09:17:08.4611054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0131s] [ 66%] 2025-07-17T09:17:08.4611519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0111s] [ 66%] 2025-07-17T09:17:08.4612034Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0117s] [ 66%] 2025-07-17T09:17:08.4612560Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0119s] [ 66%] 2025-07-17T09:17:08.4613017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0119s] [ 66%] 2025-07-17T09:17:08.4613576Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0156s] [ 66%] 2025-07-17T09:17:08.4614040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0161s] [ 66%] 2025-07-17T09:17:08.4614492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 66%] 2025-07-17T09:17:08.4614961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0050s] [ 66%] 2025-07-17T09:17:08.4615410Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0052s] [ 66%] 2025-07-17T09:17:08.4615864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0050s] [ 66%] 2025-07-17T09:17:08.4616315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0063s] [ 66%] 2025-07-17T09:17:08.4616766Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0063s] [ 66%] 2025-07-17T09:17:08.4617229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0064s] [ 66%] 2025-07-17T09:17:08.4617686Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0064s] [ 66%] 2025-07-17T09:17:08.4618138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0064s] [ 66%] 2025-07-17T09:17:08.4618591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0064s] [ 66%] 2025-07-17T09:17:08.4619119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0081s] [ 66%] 2025-07-17T09:17:08.4619624Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0080s] [ 66%] 2025-07-17T09:17:08.4620080Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 66%] 2025-07-17T09:17:08.4620654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0052s] [ 66%] 2025-07-17T09:17:08.4621106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0052s] [ 66%] 2025-07-17T09:17:08.4621567Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0052s] [ 66%] 2025-07-17T09:17:08.4622008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0061s] [ 66%] 2025-07-17T09:17:08.4622466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0062s] [ 66%] 2025-07-17T09:17:08.4622926Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0066s] [ 66%] 2025-07-17T09:17:08.4623385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0065s] [ 66%] 2025-07-17T09:17:08.4623836Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0065s] [ 66%] 2025-07-17T09:17:08.4624286Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0064s] [ 66%] 2025-07-17T09:17:08.4624741Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0080s] [ 66%] 2025-07-17T09:17:08.4625192Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0080s] [ 66%] 2025-07-17T09:17:08.4625649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0609s] [ 66%] 2025-07-17T09:17:08.4626110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0087s] [ 66%] 2025-07-17T09:17:08.4626620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0085s] [ 66%] 2025-07-17T09:17:08.4627122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0085s] [ 66%] 2025-07-17T09:17:08.4627671Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0111s] [ 66%] 2025-07-17T09:17:08.4628133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0118s] [ 66%] 2025-07-17T09:17:08.4628589Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0108s] [ 66%] 2025-07-17T09:17:08.4629055Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0112s] [ 66%] 2025-07-17T09:17:08.4629521Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0112s] [ 66%] 2025-07-17T09:17:08.4629976Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0112s] [ 66%] 2025-07-17T09:17:08.4630430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0150s] [ 66%] 2025-07-17T09:17:08.4630881Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0144s] [ 66%] 2025-07-17T09:17:08.4631343Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0148s] [ 66%] 2025-07-17T09:17:08.4631799Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0133s] [ 66%] 2025-07-17T09:17:08.4632256Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0132s] [ 66%] 2025-07-17T09:17:08.4632711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0132s] [ 66%] 2025-07-17T09:17:08.4633173Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0189s] [ 66%] 2025-07-17T09:17:08.4633705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0195s] [ 66%] 2025-07-17T09:17:08.4634275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0165s] [ 66%] 2025-07-17T09:17:08.4634814Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0169s] [ 66%] 2025-07-17T09:17:08.4635327Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0168s] [ 66%] 2025-07-17T09:17:08.4635785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0167s] [ 66%] 2025-07-17T09:17:08.4636242Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0230s] [ 66%] 2025-07-17T09:17:08.4636707Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0232s] [ 66%] 2025-07-17T09:17:08.4637164Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0111s] [ 66%] 2025-07-17T09:17:08.4637619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0080s] [ 66%] 2025-07-17T09:17:08.4638076Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0092s] [ 67%] 2025-07-17T09:17:08.4638528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0092s] [ 67%] 2025-07-17T09:17:08.4638983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0118s] [ 67%] 2025-07-17T09:17:08.4639436Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0117s] [ 67%] 2025-07-17T09:17:08.4639895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0116s] [ 67%] 2025-07-17T09:17:08.4640360Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0121s] [ 67%] 2025-07-17T09:17:08.4640890Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0122s] [ 67%] 2025-07-17T09:17:08.4641408Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0122s] [ 67%] 2025-07-17T09:17:08.4641861Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0157s] [ 67%] 2025-07-17T09:17:08.4642428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0156s] [ 67%] 2025-07-17T09:17:08.4642886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0077s] [ 67%] 2025-07-17T09:17:08.4643349Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0076s] [ 67%] 2025-07-17T09:17:08.4643802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0075s] [ 67%] 2025-07-17T09:17:08.4644254Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0075s] [ 67%] 2025-07-17T09:17:08.4644705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0100s] [ 67%] 2025-07-17T09:17:08.4645155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0100s] [ 67%] 2025-07-17T09:17:08.4645617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0101s] [ 67%] 2025-07-17T09:17:08.4646077Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0099s] [ 67%] 2025-07-17T09:17:08.4646529Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0099s] [ 67%] 2025-07-17T09:17:08.4646988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0098s] [ 67%] 2025-07-17T09:17:08.4647443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0131s] [ 67%] 2025-07-17T09:17:08.4648033Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0132s] [ 67%] 2025-07-17T09:17:08.4648540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0387s] [ 67%] 2025-07-17T09:17:08.4649000Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0121s] [ 67%] 2025-07-17T09:17:08.4649578Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0120s] [ 67%] 2025-07-17T09:17:08.4650042Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0122s] [ 67%] 2025-07-17T09:17:08.4650496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0166s] [ 67%] 2025-07-17T09:17:08.4650950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0175s] [ 67%] 2025-07-17T09:17:08.4651411Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0152s] [ 67%] 2025-07-17T09:17:08.4651869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0157s] [ 67%] 2025-07-17T09:17:08.4652321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0154s] [ 67%] 2025-07-17T09:17:08.4652774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0123s] [ 67%] 2025-07-17T09:17:08.4653227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0212s] [ 67%] 2025-07-17T09:17:08.4653678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0212s] [ 67%] 2025-07-17T09:17:08.4654137Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0088s] [ 67%] 2025-07-17T09:17:08.4654612Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0071s] [ 67%] 2025-07-17T09:17:08.4655059Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0070s] [ 67%] 2025-07-17T09:17:08.4655578Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0443s] [ 67%] 2025-07-17T09:17:08.4656076Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0646s] [ 67%] 2025-07-17T09:17:08.4656650Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.1000s] [ 67%] 2025-07-17T09:17:08.4657113Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0828s] [ 67%] 2025-07-17T09:17:08.4657583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0768s] [ 67%] 2025-07-17T09:17:08.4658037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0767s] [ 67%] 2025-07-17T09:17:08.4658493Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0733s] [ 67%] 2025-07-17T09:17:08.4658944Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0941s] [ 67%] 2025-07-17T09:17:08.4659395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0525s] [ 67%] 2025-07-17T09:17:08.4659850Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0061s] [ 67%] 2025-07-17T09:17:08.4660304Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0050s] [ 67%] 2025-07-17T09:17:08.4660758Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0050s] [ 67%] 2025-07-17T09:17:08.4661219Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0048s] [ 67%] 2025-07-17T09:17:08.4661665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0048s] [ 67%] 2025-07-17T09:17:08.4662119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0046s] [ 67%] 2025-07-17T09:17:08.4662636Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0058s] [ 67%] 2025-07-17T09:17:08.4663155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0059s] [ 67%] 2025-07-17T09:17:08.4663735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0059s] [ 67%] 2025-07-17T09:17:08.4664194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0059s] [ 67%] 2025-07-17T09:17:08.4664646Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0057s] [ 67%] 2025-07-17T09:17:08.4665101Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0057s] [ 67%] 2025-07-17T09:17:08.4665554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 67%] 2025-07-17T09:17:08.4666006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0049s] [ 67%] 2025-07-17T09:17:08.4666458Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0049s] [ 67%] 2025-07-17T09:17:08.4666903Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0050s] [ 67%] 2025-07-17T09:17:08.4667353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0047s] [ 67%] 2025-07-17T09:17:08.4667804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0047s] [ 67%] 2025-07-17T09:17:08.4668253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0059s] [ 67%] 2025-07-17T09:17:08.4668715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0056s] [ 67%] 2025-07-17T09:17:08.4669161Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0056s] [ 67%] 2025-07-17T09:17:08.4669675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0055s] [ 67%] 2025-07-17T09:17:08.4670186Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0056s] [ 67%] 2025-07-17T09:17:08.4670640Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0056s] [ 67%] 2025-07-17T09:17:08.4671216Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.7842s] [ 67%] 2025-07-17T09:17:08.4671681Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0543s] [ 67%] 2025-07-17T09:17:08.4672133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0763s] [ 67%] 2025-07-17T09:17:08.4672581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0615s] [ 67%] 2025-07-17T09:17:08.4673031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0699s] [ 67%] 2025-07-17T09:17:08.4673479Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0920s] [ 67%] 2025-07-17T09:17:08.4674011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0917s] [ 67%] 2025-07-17T09:17:08.4674475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0927s] [ 67%] 2025-07-17T09:17:08.4674928Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0870s] [ 67%] 2025-07-17T09:17:08.4675384Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0871s] [ 67%] 2025-07-17T09:17:08.4675828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0826s] [ 67%] 2025-07-17T09:17:08.4676286Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.1154s] [ 67%] 2025-07-17T09:17:08.4676808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 67%] 2025-07-17T09:17:08.4677344Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 67%] 2025-07-17T09:17:08.4677786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0047s] [ 67%] 2025-07-17T09:17:08.4678385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0047s] [ 67%] 2025-07-17T09:17:08.4678838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0050s] [ 67%] 2025-07-17T09:17:08.4679287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0048s] [ 67%] 2025-07-17T09:17:08.4679748Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 67%] 2025-07-17T09:17:08.4680204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0052s] [ 67%] 2025-07-17T09:17:08.4680656Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0052s] [ 67%] 2025-07-17T09:17:08.4681105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0052s] [ 67%] 2025-07-17T09:17:08.4681554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0053s] [ 67%] 2025-07-17T09:17:08.4682008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0053s] [ 67%] 2025-07-17T09:17:08.4682454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 67%] 2025-07-17T09:17:08.4682911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 67%] 2025-07-17T09:17:08.4683354Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 67%] 2025-07-17T09:17:08.4683807Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 67%] 2025-07-17T09:17:08.4684310Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0041s] [ 67%] 2025-07-17T09:17:08.4684820Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 67%] 2025-07-17T09:17:08.4685391Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0048s] [ 67%] 2025-07-17T09:17:08.4685843Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 67%] 2025-07-17T09:17:08.4686294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0047s] [ 67%] 2025-07-17T09:17:08.4686739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0047s] [ 67%] 2025-07-17T09:17:08.4687190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0044s] [ 67%] 2025-07-17T09:17:08.4687638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0044s] [ 67%] 2025-07-17T09:17:08.4688093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 67%] 2025-07-17T09:17:08.4688541Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 67%] 2025-07-17T09:17:08.4688994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 67%] 2025-07-17T09:17:08.4689442Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 67%] 2025-07-17T09:17:08.4689880Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0042s] [ 67%] 2025-07-17T09:17:08.4690331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 67%] 2025-07-17T09:17:08.4690778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 67%] 2025-07-17T09:17:08.4691301Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 67%] 2025-07-17T09:17:08.4691807Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0045s] [ 67%] 2025-07-17T09:17:08.4692270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 67%] 2025-07-17T09:17:08.4692854Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0043s] [ 68%] 2025-07-17T09:17:08.4693305Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 68%] 2025-07-17T09:17:08.4693757Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 68%] 2025-07-17T09:17:08.4694206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 68%] 2025-07-17T09:17:08.4694653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0040s] [ 68%] 2025-07-17T09:17:08.4695096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 68%] 2025-07-17T09:17:08.4695547Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0045s] [ 68%] 2025-07-17T09:17:08.4696007Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 68%] 2025-07-17T09:17:08.4696456Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 68%] 2025-07-17T09:17:08.4696916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 68%] 2025-07-17T09:17:08.4697358Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0045s] [ 68%] 2025-07-17T09:17:08.4697966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0049s] [ 68%] 2025-07-17T09:17:08.4704367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0052s] [ 68%] 2025-07-17T09:17:08.4704993Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0050s] [ 68%] 2025-07-17T09:17:08.4705526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.6066s] [ 68%] 2025-07-17T09:17:08.4706118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0655s] [ 68%] 2025-07-17T09:17:08.4706580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0908s] [ 68%] 2025-07-17T09:17:08.4707047Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0667s] [ 68%] 2025-07-17T09:17:08.4707495Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0861s] [ 68%] 2025-07-17T09:17:08.4707960Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.1544s] [ 68%] 2025-07-17T09:17:08.4708547Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.1299s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4709138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0021s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4709725Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0259s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4710309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4710887Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4711468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4711984Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0080s] [ 68%] 2025-07-17T09:17:08.4712499Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0607s] [ 68%] 2025-07-17T09:17:08.4712952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.1570s] [ 68%] 2025-07-17T09:17:08.4713517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0738s] [ 68%] 2025-07-17T09:17:08.4714050Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0730s] [ 68%] 2025-07-17T09:17:08.4714512Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.1168s] [ 68%] 2025-07-17T09:17:08.4715094Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.1043s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4715679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4716255Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4716832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0023s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4717405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0101s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4717980Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4718437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0084s] [ 68%] 2025-07-17T09:17:08.4718897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0081s] [ 68%] 2025-07-17T09:17:08.4719430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0081s] [ 68%] 2025-07-17T09:17:08.4719972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0080s] [ 68%] 2025-07-17T09:17:08.4720429Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0101s] [ 68%] 2025-07-17T09:17:08.4721022Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0102s] [ 68%] 2025-07-17T09:17:08.4721607Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4722188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4722754Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4723328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4723904Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4724473Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4724930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.2503s] [ 68%] 2025-07-17T09:17:08.4725397Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0786s] [ 68%] 2025-07-17T09:17:08.4725859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0573s] [ 68%] 2025-07-17T09:17:08.4726314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.1016s] [ 68%] 2025-07-17T09:17:08.4726827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0720s] [ 68%] 2025-07-17T09:17:08.4727332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.1350s] [ 68%] 2025-07-17T09:17:08.4727961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.1125s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4728596Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0021s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4729165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0045s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4729739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0337s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4730317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0083s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4730888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4731351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0189s] [ 68%] 2025-07-17T09:17:08.4731811Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1026s] [ 68%] 2025-07-17T09:17:08.4732271Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.1119s] [ 68%] 2025-07-17T09:17:08.4732724Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.1472s] [ 68%] 2025-07-17T09:17:08.4733182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0486s] [ 68%] 2025-07-17T09:17:08.4733644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0885s] [ 68%] 2025-07-17T09:17:08.4734299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0608s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4734961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0372s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4735649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4736226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4736800Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4737385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0027s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4737842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0249s] [ 68%] 2025-07-17T09:17:08.4738299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0131s] [ 68%] 2025-07-17T09:17:08.4738756Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0264s] [ 68%] 2025-07-17T09:17:08.4739208Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0389s] [ 68%] 2025-07-17T09:17:08.4739661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0458s] [ 68%] 2025-07-17T09:17:08.4740118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0720s] [ 68%] 2025-07-17T09:17:08.4740694Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0643s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4741330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4741956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4742590Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0142s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4743229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4743793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4744250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0131s] [ 68%] 2025-07-17T09:17:08.4744708Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0339s] [ 68%] 2025-07-17T09:17:08.4745162Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0365s] [ 68%] 2025-07-17T09:17:08.4745614Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0428s] [ 68%] 2025-07-17T09:17:08.4746065Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0539s] [ 68%] 2025-07-17T09:17:08.4746520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0207s] [ 68%] 2025-07-17T09:17:08.4747091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0041s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4747670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0124s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4748240Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4748890Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4749529Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4750216Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4750679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0192s] [ 68%] 2025-07-17T09:17:08.4751140Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0372s] [ 68%] 2025-07-17T09:17:08.4751599Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0567s] [ 68%] 2025-07-17T09:17:08.4752051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0364s] [ 68%] 2025-07-17T09:17:08.4752511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0516s] [ 68%] 2025-07-17T09:17:08.4752966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0750s] [ 68%] 2025-07-17T09:17:08.4753553Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0980s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4754204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0025s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4754776Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4755363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0117s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4756012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0021s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4756655Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0022s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 68%] 2025-07-17T09:17:08.4757186Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0073s] [ 68%] 2025-07-17T09:17:08.4757750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0106s] [ 68%] 2025-07-17T09:17:08.4758209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0097s] [ 68%] 2025-07-17T09:17:08.4758661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0094s] [ 68%] 2025-07-17T09:17:08.4759116Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0115s] [ 68%] 2025-07-17T09:17:08.4759570Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0279s] [ 68%] 2025-07-17T09:17:08.4760145Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0981s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4760729Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.1162s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4761293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4761868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4762439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4763006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4763515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0052s] [ 69%] 2025-07-17T09:17:08.4764019Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0048s] [ 69%] 2025-07-17T09:17:08.4764577Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0048s] [ 69%] 2025-07-17T09:17:08.4765029Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0048s] [ 69%] 2025-07-17T09:17:08.4765477Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0059s] [ 69%] 2025-07-17T09:17:08.4765932Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0057s] [ 69%] 2025-07-17T09:17:08.4766506Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4767086Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4767651Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4768236Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4768806Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4769372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4769827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0052s] [ 69%] 2025-07-17T09:17:08.4770280Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0051s] [ 69%] 2025-07-17T09:17:08.4770798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0051s] [ 69%] 2025-07-17T09:17:08.4771294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0051s] [ 69%] 2025-07-17T09:17:08.4771861Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0057s] [ 69%] 2025-07-17T09:17:08.4772319Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0058s] [ 69%] 2025-07-17T09:17:08.4772891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4773467Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4774034Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4774607Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4775177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4775746Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4776199Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.7206s] [ 69%] 2025-07-17T09:17:08.4776655Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0622s] [ 69%] 2025-07-17T09:17:08.4777109Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0820s] [ 69%] 2025-07-17T09:17:08.4777609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0864s] [ 69%] 2025-07-17T09:17:08.4778110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0871s] [ 69%] 2025-07-17T09:17:08.4778565Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0307s] [ 69%] 2025-07-17T09:17:08.4779246Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0657s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4779832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0023s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4780398Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0021s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4780972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0401s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4781536Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0023s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4782104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0109s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4782560Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0083s] [ 69%] 2025-07-17T09:17:08.4783018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0081s] [ 69%] 2025-07-17T09:17:08.4783472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0069s] [ 69%] 2025-07-17T09:17:08.4783917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0064s] [ 69%] 2025-07-17T09:17:08.4784368Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0069s] [ 69%] 2025-07-17T09:17:08.4784875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0060s] [ 69%] 2025-07-17T09:17:08.4785491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4786117Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4786736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4787314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4787878Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4788447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4788903Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 69%] 2025-07-17T09:17:08.4789355Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 69%] 2025-07-17T09:17:08.4789806Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 69%] 2025-07-17T09:17:08.4790252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 69%] 2025-07-17T09:17:08.4790699Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0043s] [ 69%] 2025-07-17T09:17:08.4791150Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 69%] 2025-07-17T09:17:08.4791717Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4792379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4792997Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4793685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4794319Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4794896Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4795355Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 69%] 2025-07-17T09:17:08.4795809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 69%] 2025-07-17T09:17:08.4796257Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 69%] 2025-07-17T09:17:08.4796703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 69%] 2025-07-17T09:17:08.4797149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0041s] [ 69%] 2025-07-17T09:17:08.4797594Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 69%] 2025-07-17T09:17:08.4798168Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4798749Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4799312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4799963Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4800580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4801299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4801754Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0048s] [ 69%] 2025-07-17T09:17:08.4802211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 69%] 2025-07-17T09:17:08.4802661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0047s] [ 69%] 2025-07-17T09:17:08.4803104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 69%] 2025-07-17T09:17:08.4803553Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0051s] [ 69%] 2025-07-17T09:17:08.4803997Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0049s] [ 69%] 2025-07-17T09:17:08.4804574Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4805149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4805711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4806281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4806910Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4807529Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 69%] 2025-07-17T09:17:08.4808040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.2426s] [ 69%] 2025-07-17T09:17:08.4808548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0073s] [ 69%] 2025-07-17T09:17:08.4809005Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0080s] [ 69%] 2025-07-17T09:17:08.4809455Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0503s] [ 69%] 2025-07-17T09:17:08.4809909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0506s] [ 69%] 2025-07-17T09:17:08.4810357Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0769s] [ 69%] 2025-07-17T09:17:08.4810824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0912s] [ 69%] 2025-07-17T09:17:08.4811288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0552s] [ 69%] 2025-07-17T09:17:08.4811738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0973s] [ 69%] 2025-07-17T09:17:08.4812194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0698s] [ 69%] 2025-07-17T09:17:08.4812641Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0525s] [ 69%] 2025-07-17T09:17:08.4813095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0907s] [ 69%] 2025-07-17T09:17:08.4813546Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0054s] [ 69%] 2025-07-17T09:17:08.4814058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0048s] [ 69%] 2025-07-17T09:17:08.4814553Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0046s] [ 69%] 2025-07-17T09:17:08.4815004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0046s] [ 69%] 2025-07-17T09:17:08.4815566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0045s] [ 69%] 2025-07-17T09:17:08.4816018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0043s] [ 69%] 2025-07-17T09:17:08.4816475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 69%] 2025-07-17T09:17:08.4816929Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0052s] [ 69%] 2025-07-17T09:17:08.4817382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0052s] [ 69%] 2025-07-17T09:17:08.4817838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0052s] [ 69%] 2025-07-17T09:17:08.4818285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0051s] [ 69%] 2025-07-17T09:17:08.4818742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0050s] [ 69%] 2025-07-17T09:17:08.4819189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 69%] 2025-07-17T09:17:08.4819647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 69%] 2025-07-17T09:17:08.4820089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0045s] [ 70%] 2025-07-17T09:17:08.4820549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 70%] 2025-07-17T09:17:08.4821048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0042s] [ 70%] 2025-07-17T09:17:08.4821541Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0043s] [ 70%] 2025-07-17T09:17:08.4821997Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0052s] [ 70%] 2025-07-17T09:17:08.4822559Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0051s] [ 70%] 2025-07-17T09:17:08.4823015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0052s] [ 70%] 2025-07-17T09:17:08.4823464Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0052s] [ 70%] 2025-07-17T09:17:08.4823911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0051s] [ 70%] 2025-07-17T09:17:08.4824361Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0051s] [ 70%] 2025-07-17T09:17:08.4824816Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [1.0102s] [ 70%] 2025-07-17T09:17:08.4825273Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0536s] [ 70%] 2025-07-17T09:17:08.4825720Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0498s] [ 70%] 2025-07-17T09:17:08.4826177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0500s] [ 70%] 2025-07-17T09:17:08.4826620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0515s] [ 70%] 2025-07-17T09:17:08.4827077Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0419s] [ 70%] 2025-07-17T09:17:08.4827531Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0937s] [ 70%] 2025-07-17T09:17:08.4827993Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0806s] [ 70%] 2025-07-17T09:17:08.4828497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0613s] [ 70%] 2025-07-17T09:17:08.4829014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0709s] [ 70%] 2025-07-17T09:17:08.4829578Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0624s] [ 70%] 2025-07-17T09:17:08.4830036Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0905s] [ 70%] 2025-07-17T09:17:08.4830496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.1209s] [ 70%] 2025-07-17T09:17:08.4830954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0789s] [ 70%] 2025-07-17T09:17:08.4831407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.1008s] [ 70%] 2025-07-17T09:17:08.4831860Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0906s] [ 70%] 2025-07-17T09:17:08.4832306Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0197s] [ 70%] 2025-07-17T09:17:08.4832762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0781s] [ 70%] 2025-07-17T09:17:08.4833217Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0929s] [ 70%] 2025-07-17T09:17:08.4833682Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1021s] [ 70%] 2025-07-17T09:17:08.4834300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0929s] [ 70%] 2025-07-17T09:17:08.4834772Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0836s] [ 70%] 2025-07-17T09:17:08.4835231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0999s] [ 70%] 2025-07-17T09:17:08.4835745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.1415s] [ 70%] 2025-07-17T09:17:08.4836265Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 70%] 2025-07-17T09:17:08.4836784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 70%] 2025-07-17T09:17:08.4837298Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0046s] [ 70%] 2025-07-17T09:17:08.4837751Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 70%] 2025-07-17T09:17:08.4838200Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0052s] [ 70%] 2025-07-17T09:17:08.4838656Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0052s] [ 70%] 2025-07-17T09:17:08.4839107Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0056s] [ 70%] 2025-07-17T09:17:08.4839571Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0055s] [ 70%] 2025-07-17T09:17:08.4840017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0055s] [ 70%] 2025-07-17T09:17:08.4840478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0056s] [ 70%] 2025-07-17T09:17:08.4840923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0066s] [ 70%] 2025-07-17T09:17:08.4841389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0065s] [ 70%] 2025-07-17T09:17:08.4841835Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 70%] 2025-07-17T09:17:08.4842297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 70%] 2025-07-17T09:17:08.4842797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0046s] [ 70%] 2025-07-17T09:17:08.4843298Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0046s] [ 70%] 2025-07-17T09:17:08.4843747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0050s] [ 70%] 2025-07-17T09:17:08.4844306Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0051s] [ 70%] 2025-07-17T09:17:08.4844764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0056s] [ 70%] 2025-07-17T09:17:08.4845221Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0056s] [ 70%] 2025-07-17T09:17:08.4845672Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0056s] [ 70%] 2025-07-17T09:17:08.4846125Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0056s] [ 70%] 2025-07-17T09:17:08.4846570Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0065s] [ 70%] 2025-07-17T09:17:08.4847027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0065s] [ 70%] 2025-07-17T09:17:08.4847474Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.9763s] [ 70%] 2025-07-17T09:17:08.4847933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0139s] [ 70%] 2025-07-17T09:17:08.4848376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0080s] [ 70%] 2025-07-17T09:17:08.4848835Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0302s] [ 70%] 2025-07-17T09:17:08.4849290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0321s] [ 70%] 2025-07-17T09:17:08.4849795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0464s] [ 70%] 2025-07-17T09:17:08.4850300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0577s] [ 70%] 2025-07-17T09:17:08.4850758Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0580s] [ 70%] 2025-07-17T09:17:08.4851318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0365s] [ 70%] 2025-07-17T09:17:08.4851771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0072s] [ 70%] 2025-07-17T09:17:08.4852222Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0458s] [ 70%] 2025-07-17T09:17:08.4852684Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0772s] [ 70%] 2025-07-17T09:17:08.4853141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0055s] [ 70%] 2025-07-17T09:17:08.4853602Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0049s] [ 70%] 2025-07-17T09:17:08.4854048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0048s] [ 70%] 2025-07-17T09:17:08.4854501Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0047s] [ 70%] 2025-07-17T09:17:08.4854947Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0050s] [ 70%] 2025-07-17T09:17:08.4855402Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0048s] [ 70%] 2025-07-17T09:17:08.4855856Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 70%] 2025-07-17T09:17:08.4856320Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0052s] [ 70%] 2025-07-17T09:17:08.4856779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0051s] [ 70%] 2025-07-17T09:17:08.4857314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0051s] [ 70%] 2025-07-17T09:17:08.4857813Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0056s] [ 70%] 2025-07-17T09:17:08.4858369Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0053s] [ 70%] 2025-07-17T09:17:08.4858832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 70%] 2025-07-17T09:17:08.4859287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 70%] 2025-07-17T09:17:08.4859734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 70%] 2025-07-17T09:17:08.4860187Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 70%] 2025-07-17T09:17:08.4860631Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0044s] [ 70%] 2025-07-17T09:17:08.4861082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 70%] 2025-07-17T09:17:08.4861529Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0051s] [ 70%] 2025-07-17T09:17:08.4861989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0049s] [ 70%] 2025-07-17T09:17:08.4862430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0049s] [ 70%] 2025-07-17T09:17:08.4862881Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0049s] [ 70%] 2025-07-17T09:17:08.4863331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0049s] [ 70%] 2025-07-17T09:17:08.4863779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0048s] [ 70%] 2025-07-17T09:17:08.4864303Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 70%] 2025-07-17T09:17:08.4864803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 70%] 2025-07-17T09:17:08.4865250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 70%] 2025-07-17T09:17:08.4865809Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 70%] 2025-07-17T09:17:08.4866260Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0041s] [ 70%] 2025-07-17T09:17:08.4866707Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 70%] 2025-07-17T09:17:08.4867162Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 70%] 2025-07-17T09:17:08.4867616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0044s] [ 70%] 2025-07-17T09:17:08.4868067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0043s] [ 70%] 2025-07-17T09:17:08.4868512Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0044s] [ 70%] 2025-07-17T09:17:08.4868968Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0046s] [ 70%] 2025-07-17T09:17:08.4869421Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0045s] [ 70%] 2025-07-17T09:17:08.4869871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 70%] 2025-07-17T09:17:08.4870324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 70%] 2025-07-17T09:17:08.4870767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 70%] 2025-07-17T09:17:08.4871291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 70%] 2025-07-17T09:17:08.4871785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 70%] 2025-07-17T09:17:08.4872235Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 70%] 2025-07-17T09:17:08.4872909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 70%] 2025-07-17T09:17:08.4873364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 70%] 2025-07-17T09:17:08.4873806Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0047s] [ 70%] 2025-07-17T09:17:08.4874336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0048s] [ 70%] 2025-07-17T09:17:08.4874791Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0048s] [ 71%] 2025-07-17T09:17:08.4875240Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0048s] [ 71%] 2025-07-17T09:17:08.4875692Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 71%] 2025-07-17T09:17:08.4876146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 71%] 2025-07-17T09:17:08.4876593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0035s] [ 71%] 2025-07-17T09:17:08.4877042Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 71%] 2025-07-17T09:17:08.4877483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4877931Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 71%] 2025-07-17T09:17:08.4878379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4878918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0051s] [ 71%] 2025-07-17T09:17:08.4879437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0043s] [ 71%] 2025-07-17T09:17:08.4880000Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 71%] 2025-07-17T09:17:08.4880561Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0044s] [ 71%] 2025-07-17T09:17:08.4881011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4881459Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 71%] 2025-07-17T09:17:08.4881906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4882348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4882794Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 71%] 2025-07-17T09:17:08.4883230Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4883683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 71%] 2025-07-17T09:17:08.4884127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 71%] 2025-07-17T09:17:08.4884581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 71%] 2025-07-17T09:17:08.4885019Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0044s] [ 71%] 2025-07-17T09:17:08.4885468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0044s] [ 71%] 2025-07-17T09:17:08.4885974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0044s] [ 71%] 2025-07-17T09:17:08.4886485Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 71%] 2025-07-17T09:17:08.4886931Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 71%] 2025-07-17T09:17:08.4887509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4887957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 71%] 2025-07-17T09:17:08.4888402Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4888842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 71%] 2025-07-17T09:17:08.4889285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 71%] 2025-07-17T09:17:08.4889733Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 71%] 2025-07-17T09:17:08.4890183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4890618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0042s] [ 71%] 2025-07-17T09:17:08.4891064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4891500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 71%] 2025-07-17T09:17:08.4891946Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 71%] 2025-07-17T09:17:08.4892386Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4892840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4893344Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 71%] 2025-07-17T09:17:08.4893845Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 71%] 2025-07-17T09:17:08.4894435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 71%] 2025-07-17T09:17:08.4894885Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 71%] 2025-07-17T09:17:08.4895341Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4895791Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4896245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0042s] [ 71%] 2025-07-17T09:17:08.4896686Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4897127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0039s] [ 71%] 2025-07-17T09:17:08.4897571Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_312_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 71%] 2025-07-17T09:17:08.4898031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4898490Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 71%] 2025-07-17T09:17:08.4898932Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4899379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 71%] 2025-07-17T09:17:08.4899822Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4900329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 71%] 2025-07-17T09:17:08.4900850Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 71%] 2025-07-17T09:17:08.4901304Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 71%] 2025-07-17T09:17:08.4901883Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0046s] [ 71%] 2025-07-17T09:17:08.4902338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0046s] [ 71%] 2025-07-17T09:17:08.4902785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0045s] [ 71%] 2025-07-17T09:17:08.4903236Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0045s] [ 71%] 2025-07-17T09:17:08.4903686Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 71%] 2025-07-17T09:17:08.4904137Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4904583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 71%] 2025-07-17T09:17:08.4905032Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 71%] 2025-07-17T09:17:08.4905472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4905921Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 71%] 2025-07-17T09:17:08.4906369Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4906827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4907343Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4907862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 71%] 2025-07-17T09:17:08.4908309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 71%] 2025-07-17T09:17:08.4908914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 71%] 2025-07-17T09:17:08.4909368Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 71%] 2025-07-17T09:17:08.4909817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4910254Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4910698Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4911138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 71%] 2025-07-17T09:17:08.4911583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 71%] 2025-07-17T09:17:08.4912030Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0057s] [ 71%] 2025-07-17T09:17:08.4912487Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 71%] 2025-07-17T09:17:08.4912926Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0048s] [ 71%] 2025-07-17T09:17:08.4913375Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 71%] 2025-07-17T09:17:08.4913813Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0042s] [ 71%] 2025-07-17T09:17:08.4914332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0043s] [ 71%] 2025-07-17T09:17:08.4914873Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4915388Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 71%] 2025-07-17T09:17:08.4915908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 71%] 2025-07-17T09:17:08.4916426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 71%] 2025-07-17T09:17:08.4916878Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0043s] [ 71%] 2025-07-17T09:17:08.4917331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4917781Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 71%] 2025-07-17T09:17:08.4918235Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 71%] 2025-07-17T09:17:08.4918677Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0045s] [ 71%] 2025-07-17T09:17:08.4919129Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0044s] [ 71%] 2025-07-17T09:17:08.4919569Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0047s] [ 71%] 2025-07-17T09:17:08.4920020Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0054s] [ 71%] 2025-07-17T09:17:08.4920469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 71%] 2025-07-17T09:17:08.4920920Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 71%] 2025-07-17T09:17:08.4921370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0043s] [ 71%] 2025-07-17T09:17:08.4921884Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 71%] 2025-07-17T09:17:08.4922387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0050s] [ 71%] 2025-07-17T09:17:08.4922831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0048s] [ 71%] 2025-07-17T09:17:08.4923431Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0050s] [ 71%] 2025-07-17T09:17:08.4923900Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0047s] [ 71%] 2025-07-17T09:17:08.4924347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0051s] [ 71%] 2025-07-17T09:17:08.4924790Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0049s] [ 71%] 2025-07-17T09:17:08.4925238Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0062s] [ 71%] 2025-07-17T09:17:08.4925687Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0052s] [ 71%] 2025-07-17T09:17:08.4926131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4926585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 71%] 2025-07-17T09:17:08.4927024Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0035s] [ 71%] 2025-07-17T09:17:08.4927471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 71%] 2025-07-17T09:17:08.4927917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 71%] 2025-07-17T09:17:08.4928366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 71%] 2025-07-17T09:17:08.4928815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 72%] 2025-07-17T09:17:08.4929330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0044s] [ 72%] 2025-07-17T09:17:08.4929836Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0042s] [ 72%] 2025-07-17T09:17:08.4930408Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 72%] 2025-07-17T09:17:08.4930855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0043s] [ 72%] 2025-07-17T09:17:08.4931299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0043s] [ 72%] 2025-07-17T09:17:08.4931755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4932212Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 72%] 2025-07-17T09:17:08.4932653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 72%] 2025-07-17T09:17:08.4933100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4933536Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4933981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 72%] 2025-07-17T09:17:08.4934425Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 72%] 2025-07-17T09:17:08.4934881Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 72%] 2025-07-17T09:17:08.4935330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0042s] [ 72%] 2025-07-17T09:17:08.4935775Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4936277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0039s] [ 72%] 2025-07-17T09:17:08.4936796Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4937245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4937824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4938275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4938720Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 72%] 2025-07-17T09:17:08.4939164Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0046s] [ 72%] 2025-07-17T09:17:08.4939613Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0044s] [ 72%] 2025-07-17T09:17:08.4940060Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 72%] 2025-07-17T09:17:08.4940516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 72%] 2025-07-17T09:17:08.4940956Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0043s] [ 72%] 2025-07-17T09:17:08.4941403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0046s] [ 72%] 2025-07-17T09:17:08.4941846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0051s] [ 72%] 2025-07-17T09:17:08.4942296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0043s] [ 72%] 2025-07-17T09:17:08.4942746Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 72%] 2025-07-17T09:17:08.4943269Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 72%] 2025-07-17T09:17:08.4943772Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 72%] 2025-07-17T09:17:08.4944216Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 72%] 2025-07-17T09:17:08.4944783Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0045s] [ 72%] 2025-07-17T09:17:08.4945234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 72%] 2025-07-17T09:17:08.4945691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 72%] 2025-07-17T09:17:08.4946143Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0048s] [ 72%] 2025-07-17T09:17:08.4946583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0048s] [ 72%] 2025-07-17T09:17:08.4947033Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 72%] 2025-07-17T09:17:08.4947474Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0043s] [ 72%] 2025-07-17T09:17:08.4947921Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4948363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4948814Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 72%] 2025-07-17T09:17:08.4949260Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 72%] 2025-07-17T09:17:08.4949700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 72%] 2025-07-17T09:17:08.4950146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 72%] 2025-07-17T09:17:08.4950658Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 72%] 2025-07-17T09:17:08.4951165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 72%] 2025-07-17T09:17:08.4951676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 72%] 2025-07-17T09:17:08.4952181Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4952628Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4953073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0039s] [ 72%] 2025-07-17T09:17:08.4953519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4954032Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 72%] 2025-07-17T09:17:08.4954482Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 72%] 2025-07-17T09:17:08.4954927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 72%] 2025-07-17T09:17:08.4955374Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 72%] 2025-07-17T09:17:08.4955807Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 72%] 2025-07-17T09:17:08.4956255Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 72%] 2025-07-17T09:17:08.4956700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4957152Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4957676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 72%] 2025-07-17T09:17:08.4958193Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4958637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 72%] 2025-07-17T09:17:08.4959228Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 72%] 2025-07-17T09:17:08.4959677Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4960128Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 72%] 2025-07-17T09:17:08.4960563Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0035s] [ 72%] 2025-07-17T09:17:08.4961008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 72%] 2025-07-17T09:17:08.4961445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 72%] 2025-07-17T09:17:08.4961891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 72%] 2025-07-17T09:17:08.4962339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 72%] 2025-07-17T09:17:08.4962793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4963228Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4963678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4964122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4964565Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_408_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 72%] 2025-07-17T09:17:08.4965095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 72%] 2025-07-17T09:17:08.4965596Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 72%] 2025-07-17T09:17:08.4966100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 72%] 2025-07-17T09:17:08.4966597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 72%] 2025-07-17T09:17:08.4967038Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 72%] 2025-07-17T09:17:08.4967482Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 72%] 2025-07-17T09:17:08.4967929Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4968376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 72%] 2025-07-17T09:17:08.4968812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 72%] 2025-07-17T09:17:08.4969255Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 72%] 2025-07-17T09:17:08.4969691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 72%] 2025-07-17T09:17:08.4970138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4970580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 72%] 2025-07-17T09:17:08.4971022Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 72%] 2025-07-17T09:17:08.4971462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 72%] 2025-07-17T09:17:08.4971967Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 72%] 2025-07-17T09:17:08.4972468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 72%] 2025-07-17T09:17:08.4972909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0107s] [ 72%] 2025-07-17T09:17:08.4973479Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0088s] [ 72%] 2025-07-17T09:17:08.4973938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0055s] [ 72%] 2025-07-17T09:17:08.4974376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0047s] [ 72%] 2025-07-17T09:17:08.4974821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 72%] 2025-07-17T09:17:08.4975258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0045s] [ 72%] 2025-07-17T09:17:08.4975704Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0043s] [ 72%] 2025-07-17T09:17:08.4976143Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 72%] 2025-07-17T09:17:08.4976588Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 72%] 2025-07-17T09:17:08.4977026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0039s] [ 72%] 2025-07-17T09:17:08.4977471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 72%] 2025-07-17T09:17:08.4977908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4978345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 72%] 2025-07-17T09:17:08.4978792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 72%] 2025-07-17T09:17:08.4979298Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 72%] 2025-07-17T09:17:08.4979805Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0041s] [ 72%] 2025-07-17T09:17:08.4980308Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4980813Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 72%] 2025-07-17T09:17:08.4981256Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 72%] 2025-07-17T09:17:08.4981694Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 72%] 2025-07-17T09:17:08.4982146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 72%] 2025-07-17T09:17:08.4982581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 73%] 2025-07-17T09:17:08.4983021Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 73%] 2025-07-17T09:17:08.4983456Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 73%] 2025-07-17T09:17:08.4983909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 73%] 2025-07-17T09:17:08.4984355Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 73%] 2025-07-17T09:17:08.4984810Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 73%] 2025-07-17T09:17:08.4985263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0042s] [ 73%] 2025-07-17T09:17:08.4985704Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 73%] 2025-07-17T09:17:08.4986142Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0043s] [ 73%] 2025-07-17T09:17:08.4986649Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_attn_mask_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 73%] 2025-07-17T09:17:08.4987059Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_mask_variants_mask_dim_1_cuda PASSED [0.0012s] [ 73%] 2025-07-17T09:17:08.4987315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_mask_variants_mask_dim_2_cuda PASSED [0.0011s] [ 73%] 2025-07-17T09:17:08.4987688Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_mask_variants_mask_dim_3_cuda PASSED [0.0011s] [ 73%] 2025-07-17T09:17:08.4987941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_mask_variants_mask_dim_4_cuda PASSED [0.0011s] [ 73%] 2025-07-17T09:17:08.4988387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0172s] [ 73%] 2025-07-17T09:17:08.4988838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 73%] 2025-07-17T09:17:08.4989272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0149s] [ 73%] 2025-07-17T09:17:08.4989718Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0051s] [ 73%] 2025-07-17T09:17:08.4990149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0120s] [ 73%] 2025-07-17T09:17:08.4990587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 73%] 2025-07-17T09:17:08.4991022Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0157s] [ 73%] 2025-07-17T09:17:08.4991473Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 73%] 2025-07-17T09:17:08.4991903Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0156s] [ 73%] 2025-07-17T09:17:08.4992343Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 73%] 2025-07-17T09:17:08.4992778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0140s] [ 73%] 2025-07-17T09:17:08.4993216Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 73%] 2025-07-17T09:17:08.4993703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0160s] [ 73%] 2025-07-17T09:17:08.4994247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 73%] 2025-07-17T09:17:08.4994737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0143s] [ 73%] 2025-07-17T09:17:08.4995228Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 73%] 2025-07-17T09:17:08.4995660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0107s] [ 73%] 2025-07-17T09:17:08.4996093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 73%] 2025-07-17T09:17:08.4996526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0171s] [ 73%] 2025-07-17T09:17:08.4996981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0042s] [ 73%] 2025-07-17T09:17:08.4997413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0164s] [ 73%] 2025-07-17T09:17:08.4997849Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0043s] [ 73%] 2025-07-17T09:17:08.4998276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0125s] [ 73%] 2025-07-17T09:17:08.4998711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0043s] [ 73%] 2025-07-17T09:17:08.4999145Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0124s] [ 73%] 2025-07-17T09:17:08.4999587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 73%] 2025-07-17T09:17:08.5000017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0115s] [ 73%] 2025-07-17T09:17:08.5000449Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 73%] 2025-07-17T09:17:08.5000936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0086s] [ 73%] 2025-07-17T09:17:08.5001434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 73%] 2025-07-17T09:17:08.5001986Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0156s] [ 73%] 2025-07-17T09:17:08.5002427Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 73%] 2025-07-17T09:17:08.5002862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0147s] [ 73%] 2025-07-17T09:17:08.5003299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 73%] 2025-07-17T09:17:08.5003729Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0108s] [ 73%] 2025-07-17T09:17:08.5004161Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 73%] 2025-07-17T09:17:08.5004592Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0127s] [ 73%] 2025-07-17T09:17:08.5005031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 73%] 2025-07-17T09:17:08.5005457Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0113s] [ 73%] 2025-07-17T09:17:08.5005891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 73%] 2025-07-17T09:17:08.5006318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0097s] [ 73%] 2025-07-17T09:17:08.5006755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 73%] 2025-07-17T09:17:08.5007195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0165s] [ 73%] 2025-07-17T09:17:08.5007638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 73%] 2025-07-17T09:17:08.5008150Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0165s] [ 73%] 2025-07-17T09:17:08.5008630Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0047s] [ 73%] 2025-07-17T09:17:08.5009165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0107s] [ 73%] 2025-07-17T09:17:08.5009598Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 73%] 2025-07-17T09:17:08.5010037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 73%] 2025-07-17T09:17:08.5010473Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 73%] 2025-07-17T09:17:08.5010904Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 73%] 2025-07-17T09:17:08.5011336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 73%] 2025-07-17T09:17:08.5011763Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0090s] [ 73%] 2025-07-17T09:17:08.5012197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 73%] 2025-07-17T09:17:08.5012628Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 73%] 2025-07-17T09:17:08.5013072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 73%] 2025-07-17T09:17:08.5013497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 73%] 2025-07-17T09:17:08.5013932Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 73%] 2025-07-17T09:17:08.5014365Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0101s] [ 73%] 2025-07-17T09:17:08.5014854Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 73%] 2025-07-17T09:17:08.5015336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0031s] [ 73%] 2025-07-17T09:17:08.5015782Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 73%] 2025-07-17T09:17:08.5016333Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 73%] 2025-07-17T09:17:08.5016763Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 73%] 2025-07-17T09:17:08.5017195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0092s] [ 73%] 2025-07-17T09:17:08.5017631Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 73%] 2025-07-17T09:17:08.5018064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 73%] 2025-07-17T09:17:08.5018509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 73%] 2025-07-17T09:17:08.5018933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 73%] 2025-07-17T09:17:08.5019366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 73%] 2025-07-17T09:17:08.5019795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0100s] [ 73%] 2025-07-17T09:17:08.5020233Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 73%] 2025-07-17T09:17:08.5020667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0144s] [ 73%] 2025-07-17T09:17:08.5021113Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 73%] 2025-07-17T09:17:08.5021551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0134s] [ 73%] 2025-07-17T09:17:08.5022050Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 73%] 2025-07-17T09:17:08.5022536Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0094s] [ 73%] 2025-07-17T09:17:08.5022973Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 73%] 2025-07-17T09:17:08.5023518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0192s] [ 73%] 2025-07-17T09:17:08.5023961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0047s] [ 73%] 2025-07-17T09:17:08.5024395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0215s] [ 73%] 2025-07-17T09:17:08.5024832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0052s] [ 73%] 2025-07-17T09:17:08.5025264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0128s] [ 73%] 2025-07-17T09:17:08.5025701Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 73%] 2025-07-17T09:17:08.5026133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0168s] [ 73%] 2025-07-17T09:17:08.5026571Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 73%] 2025-07-17T09:17:08.5026996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0148s] [ 73%] 2025-07-17T09:17:08.5027437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 73%] 2025-07-17T09:17:08.5027862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0103s] [ 73%] 2025-07-17T09:17:08.5028301Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 73%] 2025-07-17T09:17:08.5028743Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0168s] [ 73%] 2025-07-17T09:17:08.5029232Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 73%] 2025-07-17T09:17:08.5029714Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0155s] [ 73%] 2025-07-17T09:17:08.5030145Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 73%] 2025-07-17T09:17:08.5030682Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0107s] [ 73%] 2025-07-17T09:17:08.5031119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 73%] 2025-07-17T09:17:08.5031562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0054s] [ 73%] 2025-07-17T09:17:08.5032004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 73%] 2025-07-17T09:17:08.5032434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0049s] [ 73%] 2025-07-17T09:17:08.5032877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 73%] 2025-07-17T09:17:08.5033304Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0041s] [ 73%] 2025-07-17T09:17:08.5033744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 73%] 2025-07-17T09:17:08.5034252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 74%] 2025-07-17T09:17:08.5034699Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 74%] 2025-07-17T09:17:08.5035127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0052s] [ 74%] 2025-07-17T09:17:08.5035565Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 74%] 2025-07-17T09:17:08.5036011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0053s] [ 74%] 2025-07-17T09:17:08.5036508Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0048s] [ 74%] 2025-07-17T09:17:08.5037000Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0076s] [ 74%] 2025-07-17T09:17:08.5037498Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0047s] [ 74%] 2025-07-17T09:17:08.5037987Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0084s] [ 74%] 2025-07-17T09:17:08.5038416Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0048s] [ 74%] 2025-07-17T09:17:08.5038852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0045s] [ 74%] 2025-07-17T09:17:08.5039295Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0044s] [ 74%] 2025-07-17T09:17:08.5039728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 74%] 2025-07-17T09:17:08.5040169Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0051s] [ 74%] 2025-07-17T09:17:08.5040594Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0041s] [ 74%] 2025-07-17T09:17:08.5041031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 74%] 2025-07-17T09:17:08.5041456Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0048s] [ 74%] 2025-07-17T09:17:08.5041893Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 74%] 2025-07-17T09:17:08.5042330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 74%] 2025-07-17T09:17:08.5042779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 74%] 2025-07-17T09:17:08.5043210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0035s] [ 74%] 2025-07-17T09:17:08.5043687Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 74%] 2025-07-17T09:17:08.5044183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 74%] 2025-07-17T09:17:08.5044662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 74%] 2025-07-17T09:17:08.5045150Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 74%] 2025-07-17T09:17:08.5045588Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 74%] 2025-07-17T09:17:08.5046021Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0046s] [ 74%] 2025-07-17T09:17:08.5046465Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 74%] 2025-07-17T09:17:08.5046895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0046s] [ 74%] 2025-07-17T09:17:08.5047329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 74%] 2025-07-17T09:17:08.5047764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 74%] 2025-07-17T09:17:08.5048200Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 74%] 2025-07-17T09:17:08.5048625Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 74%] 2025-07-17T09:17:08.5049056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 74%] 2025-07-17T09:17:08.5049483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 74%] 2025-07-17T09:17:08.5049909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 74%] 2025-07-17T09:17:08.5050346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 74%] 2025-07-17T09:17:08.5050836Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 74%] 2025-07-17T09:17:08.5051310Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0059s] [ 74%] 2025-07-17T09:17:08.5051851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 74%] 2025-07-17T09:17:08.5052287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0054s] [ 74%] 2025-07-17T09:17:08.5052716Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 74%] 2025-07-17T09:17:08.5053152Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0052s] [ 74%] 2025-07-17T09:17:08.5053592Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 74%] 2025-07-17T09:17:08.5054016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0047s] [ 74%] 2025-07-17T09:17:08.5054451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 74%] 2025-07-17T09:17:08.5054875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0051s] [ 74%] 2025-07-17T09:17:08.5055313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 74%] 2025-07-17T09:17:08.5055740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0051s] [ 74%] 2025-07-17T09:17:08.5056184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 74%] 2025-07-17T09:17:08.5056613Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0048s] [ 74%] 2025-07-17T09:17:08.5057049Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 74%] 2025-07-17T09:17:08.5057478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0062s] [ 74%] 2025-07-17T09:17:08.5057960Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 74%] 2025-07-17T09:17:08.5058441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0064s] [ 74%] 2025-07-17T09:17:08.5058974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 74%] 2025-07-17T09:17:08.5059407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0061s] [ 74%] 2025-07-17T09:17:08.5059833Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 74%] 2025-07-17T09:17:08.5060263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0043s] [ 74%] 2025-07-17T09:17:08.5060694Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 74%] 2025-07-17T09:17:08.5061120Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0073s] [ 74%] 2025-07-17T09:17:08.5061555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 74%] 2025-07-17T09:17:08.5061978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0074s] [ 74%] 2025-07-17T09:17:08.5062409Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 74%] 2025-07-17T09:17:08.5062831Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0044s] [ 74%] 2025-07-17T09:17:08.5063263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 74%] 2025-07-17T09:17:08.5063696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 74%] 2025-07-17T09:17:08.5064137Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 74%] 2025-07-17T09:17:08.5064565Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0045s] [ 74%] 2025-07-17T09:17:08.5065046Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 74%] 2025-07-17T09:17:08.5065539Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0047s] [ 74%] 2025-07-17T09:17:08.5066082Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 74%] 2025-07-17T09:17:08.5066528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 74%] 2025-07-17T09:17:08.5066965Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 74%] 2025-07-17T09:17:08.5067397Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0049s] [ 74%] 2025-07-17T09:17:08.5067834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 74%] 2025-07-17T09:17:08.5068258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0047s] [ 74%] 2025-07-17T09:17:08.5068696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 74%] 2025-07-17T09:17:08.5069120Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 74%] 2025-07-17T09:17:08.5069562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 74%] 2025-07-17T09:17:08.5069985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0048s] [ 74%] 2025-07-17T09:17:08.5070416Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 74%] 2025-07-17T09:17:08.5070837Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0045s] [ 74%] 2025-07-17T09:17:08.5071267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 74%] 2025-07-17T09:17:08.5071764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0077s] [ 74%] 2025-07-17T09:17:08.5072244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 74%] 2025-07-17T09:17:08.5072676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0071s] [ 74%] 2025-07-17T09:17:08.5073211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 74%] 2025-07-17T09:17:08.5073644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0048s] [ 74%] 2025-07-17T09:17:08.5074142Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 74%] 2025-07-17T09:17:08.5074585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.1147s] [ 74%] 2025-07-17T09:17:08.5075039Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0067s] [ 74%] 2025-07-17T09:17:08.5075471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0087s] [ 74%] 2025-07-17T09:17:08.5075909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0077s] [ 74%] 2025-07-17T09:17:08.5076334Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0084s] [ 74%] 2025-07-17T09:17:08.5076780Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0066s] [ 74%] 2025-07-17T09:17:08.5077219Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0099s] [ 74%] 2025-07-17T09:17:08.5077680Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0084s] [ 74%] 2025-07-17T09:17:08.5078114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0098s] [ 74%] 2025-07-17T09:17:08.5078553Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0076s] [ 74%] 2025-07-17T09:17:08.5079054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0077s] [ 74%] 2025-07-17T09:17:08.5079546Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0072s] [ 74%] 2025-07-17T09:17:08.5079986Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0063s] [ 74%] 2025-07-17T09:17:08.5080550Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0060s] [ 74%] 2025-07-17T09:17:08.5080992Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0060s] [ 74%] 2025-07-17T09:17:08.5081428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0059s] [ 74%] 2025-07-17T09:17:08.5081864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0061s] [ 74%] 2025-07-17T09:17:08.5082299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0056s] [ 74%] 2025-07-17T09:17:08.5082732Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0064s] [ 74%] 2025-07-17T09:17:08.5083177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0074s] [ 74%] 2025-07-17T09:17:08.5083603Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0075s] [ 74%] 2025-07-17T09:17:08.5084040Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0074s] [ 74%] 2025-07-17T09:17:08.5084469Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0076s] [ 74%] 2025-07-17T09:17:08.5084910Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0070s] [ 74%] 2025-07-17T09:17:08.5085346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 74%] 2025-07-17T09:17:08.5085785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 74%] 2025-07-17T09:17:08.5086271Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 75%] 2025-07-17T09:17:08.5086750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 75%] 2025-07-17T09:17:08.5087180Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 75%] 2025-07-17T09:17:08.5087717Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 75%] 2025-07-17T09:17:08.5088161Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 75%] 2025-07-17T09:17:08.5088612Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 75%] 2025-07-17T09:17:08.5089051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0045s] [ 75%] 2025-07-17T09:17:08.5089492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 75%] 2025-07-17T09:17:08.5089920Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0044s] [ 75%] 2025-07-17T09:17:08.5090363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0044s] [ 75%] 2025-07-17T09:17:08.5090792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 75%] 2025-07-17T09:17:08.5091235Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 75%] 2025-07-17T09:17:08.5091661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0039s] [ 75%] 2025-07-17T09:17:08.5092096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 75%] 2025-07-17T09:17:08.5092526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 75%] 2025-07-17T09:17:08.5092955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 75%] 2025-07-17T09:17:08.5093443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0048s] [ 75%] 2025-07-17T09:17:08.5093926Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0047s] [ 75%] 2025-07-17T09:17:08.5094407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0046s] [ 75%] 2025-07-17T09:17:08.5094901Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0046s] [ 75%] 2025-07-17T09:17:08.5095339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0045s] [ 75%] 2025-07-17T09:17:08.5095772Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0045s] [ 75%] 2025-07-17T09:17:08.5096219Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 75%] 2025-07-17T09:17:08.5096665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 75%] 2025-07-17T09:17:08.5097091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 75%] 2025-07-17T09:17:08.5097532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 75%] 2025-07-17T09:17:08.5097961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0043s] [ 75%] 2025-07-17T09:17:08.5098397Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 75%] 2025-07-17T09:17:08.5098828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 75%] 2025-07-17T09:17:08.5099270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 75%] 2025-07-17T09:17:08.5099703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 75%] 2025-07-17T09:17:08.5100137Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 75%] 2025-07-17T09:17:08.5100760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 75%] 2025-07-17T09:17:08.5101250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 75%] 2025-07-17T09:17:08.5101745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 75%] 2025-07-17T09:17:08.5102237Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 75%] 2025-07-17T09:17:08.5102673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 75%] 2025-07-17T09:17:08.5103100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 75%] 2025-07-17T09:17:08.5103533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 75%] 2025-07-17T09:17:08.5103972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 75%] 2025-07-17T09:17:08.5104401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 75%] 2025-07-17T09:17:08.5104846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 75%] 2025-07-17T09:17:08.5105276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0043s] [ 75%] 2025-07-17T09:17:08.5105709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 75%] 2025-07-17T09:17:08.5106132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0041s] [ 75%] 2025-07-17T09:17:08.5106566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 75%] 2025-07-17T09:17:08.5107006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0689s] [ 75%] 2025-07-17T09:17:08.5107443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0059s] [ 75%] 2025-07-17T09:17:08.5107929Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0063s] [ 75%] 2025-07-17T09:17:08.5108412Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0055s] [ 75%] 2025-07-17T09:17:08.5108949Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0067s] [ 75%] 2025-07-17T09:17:08.5109381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0057s] [ 75%] 2025-07-17T09:17:08.5109822Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0068s] [ 75%] 2025-07-17T09:17:08.5110264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0069s] [ 75%] 2025-07-17T09:17:08.5110702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0075s] [ 75%] 2025-07-17T09:17:08.5111136Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0068s] [ 75%] 2025-07-17T09:17:08.5111570Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0079s] [ 75%] 2025-07-17T09:17:08.5112009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0069s] [ 75%] 2025-07-17T09:17:08.5112446Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0060s] [ 75%] 2025-07-17T09:17:08.5112885Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0057s] [ 75%] 2025-07-17T09:17:08.5113310Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0058s] [ 75%] 2025-07-17T09:17:08.5113748Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0057s] [ 75%] 2025-07-17T09:17:08.5114248Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0061s] [ 75%] 2025-07-17T09:17:08.5114678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 75%] 2025-07-17T09:17:08.5115174Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0062s] [ 75%] 2025-07-17T09:17:08.5115670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0069s] [ 75%] 2025-07-17T09:17:08.5116226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0070s] [ 75%] 2025-07-17T09:17:08.5116658Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0070s] [ 75%] 2025-07-17T09:17:08.5117097Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0070s] [ 75%] 2025-07-17T09:17:08.5117526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0066s] [ 75%] 2025-07-17T09:17:08.5117961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0060s] [ 75%] 2025-07-17T09:17:08.5118399Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 75%] 2025-07-17T09:17:08.5118825Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0055s] [ 75%] 2025-07-17T09:17:08.5119259Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 75%] 2025-07-17T09:17:08.5119685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0045s] [ 75%] 2025-07-17T09:17:08.5120120Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 75%] 2025-07-17T09:17:08.5120549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0056s] [ 75%] 2025-07-17T09:17:08.5120988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 75%] 2025-07-17T09:17:08.5121419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0056s] [ 75%] 2025-07-17T09:17:08.5121912Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 75%] 2025-07-17T09:17:08.5122395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0054s] [ 75%] 2025-07-17T09:17:08.5122818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 75%] 2025-07-17T09:17:08.5123396Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 75%] 2025-07-17T09:17:08.5123832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 75%] 2025-07-17T09:17:08.5124262Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0047s] [ 75%] 2025-07-17T09:17:08.5124685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 75%] 2025-07-17T09:17:08.5125112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0050s] [ 75%] 2025-07-17T09:17:08.5125544Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 75%] 2025-07-17T09:17:08.5125971Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0057s] [ 75%] 2025-07-17T09:17:08.5126406Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 75%] 2025-07-17T09:17:08.5126832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0060s] [ 75%] 2025-07-17T09:17:08.5127261Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 75%] 2025-07-17T09:17:08.5127683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0055s] [ 75%] 2025-07-17T09:17:08.5128114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 75%] 2025-07-17T09:17:08.5128544Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 75%] 2025-07-17T09:17:08.5129031Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 75%] 2025-07-17T09:17:08.5129514Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0044s] [ 75%] 2025-07-17T09:17:08.5129937Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 75%] 2025-07-17T09:17:08.5130466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0042s] [ 75%] 2025-07-17T09:17:08.5130899Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 75%] 2025-07-17T09:17:08.5131336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 75%] 2025-07-17T09:17:08.5131769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 75%] 2025-07-17T09:17:08.5132199Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0047s] [ 75%] 2025-07-17T09:17:08.5132630Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 75%] 2025-07-17T09:17:08.5133052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0045s] [ 75%] 2025-07-17T09:17:08.5133481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 75%] 2025-07-17T09:17:08.5133905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 75%] 2025-07-17T09:17:08.5134337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 75%] 2025-07-17T09:17:08.5134759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0043s] [ 75%] 2025-07-17T09:17:08.5135190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 75%] 2025-07-17T09:17:08.5135613Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 75%] 2025-07-17T09:17:08.5136088Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 75%] 2025-07-17T09:17:08.5136571Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0054s] [ 75%] 2025-07-17T09:17:08.5136999Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0044s] [ 75%] 2025-07-17T09:17:08.5137549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0052s] [ 75%] 2025-07-17T09:17:08.5137978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 75%] 2025-07-17T09:17:08.5138408Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0062s] [ 76%] 2025-07-17T09:17:08.5138832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 76%] 2025-07-17T09:17:08.5139267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 76%] 2025-07-17T09:17:08.5139700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 76%] 2025-07-17T09:17:08.5140121Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0040s] [ 76%] 2025-07-17T09:17:08.5140550Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 76%] 2025-07-17T09:17:08.5140971Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0051s] [ 76%] 2025-07-17T09:17:08.5141403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 76%] 2025-07-17T09:17:08.5141830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 76%] 2025-07-17T09:17:08.5142268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 76%] 2025-07-17T09:17:08.5142692Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 76%] 2025-07-17T09:17:08.5143185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 76%] 2025-07-17T09:17:08.5143661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0050s] [ 76%] 2025-07-17T09:17:08.5144084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 76%] 2025-07-17T09:17:08.5144623Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0055s] [ 76%] 2025-07-17T09:17:08.5145054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 76%] 2025-07-17T09:17:08.5145482Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0052s] [ 76%] 2025-07-17T09:17:08.5145905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 76%] 2025-07-17T09:17:08.5146333Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 76%] 2025-07-17T09:17:08.5146757Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 76%] 2025-07-17T09:17:08.5147184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 76%] 2025-07-17T09:17:08.5147615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 76%] 2025-07-17T09:17:08.5148035Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0042s] [ 76%] 2025-07-17T09:17:08.5148461Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 76%] 2025-07-17T09:17:08.5148882Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0050s] [ 76%] 2025-07-17T09:17:08.5149313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 76%] 2025-07-17T09:17:08.5149748Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0048s] [ 76%] 2025-07-17T09:17:08.5150186Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 76%] 2025-07-17T09:17:08.5150672Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0051s] [ 76%] 2025-07-17T09:17:08.5151172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 76%] 2025-07-17T09:17:08.5151705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 76%] 2025-07-17T09:17:08.5152134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 76%] 2025-07-17T09:17:08.5152572Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0055s] [ 76%] 2025-07-17T09:17:08.5153004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 76%] 2025-07-17T09:17:08.5153434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0056s] [ 76%] 2025-07-17T09:17:08.5153871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 76%] 2025-07-17T09:17:08.5154364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 76%] 2025-07-17T09:17:08.5154796Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 76%] 2025-07-17T09:17:08.5155220Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0051s] [ 76%] 2025-07-17T09:17:08.5155654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 76%] 2025-07-17T09:17:08.5156075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0051s] [ 76%] 2025-07-17T09:17:08.5156502Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 76%] 2025-07-17T09:17:08.5156923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0044s] [ 76%] 2025-07-17T09:17:08.5157355Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 76%] 2025-07-17T09:17:08.5157858Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0050s] [ 76%] 2025-07-17T09:17:08.5158359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 76%] 2025-07-17T09:17:08.5158910Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0050s] [ 76%] 2025-07-17T09:17:08.5159335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 76%] 2025-07-17T09:17:08.5159759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0049s] [ 76%] 2025-07-17T09:17:08.5160183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 76%] 2025-07-17T09:17:08.5160629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0089s] [ 76%] 2025-07-17T09:17:08.5161074Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 76%] 2025-07-17T09:17:08.5161502Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0077s] [ 76%] 2025-07-17T09:17:08.5161947Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 76%] 2025-07-17T09:17:08.5162376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0072s] [ 76%] 2025-07-17T09:17:08.5162815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 76%] 2025-07-17T09:17:08.5163250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0097s] [ 76%] 2025-07-17T09:17:08.5163699Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 76%] 2025-07-17T09:17:08.5164127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0092s] [ 76%] 2025-07-17T09:17:08.5164624Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 76%] 2025-07-17T09:17:08.5165112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0077s] [ 76%] 2025-07-17T09:17:08.5165545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 76%] 2025-07-17T09:17:08.5166097Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0084s] [ 76%] 2025-07-17T09:17:08.5166535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 76%] 2025-07-17T09:17:08.5166968Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0078s] [ 76%] 2025-07-17T09:17:08.5167395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 76%] 2025-07-17T09:17:08.5167830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0067s] [ 76%] 2025-07-17T09:17:08.5168263Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 76%] 2025-07-17T09:17:08.5168697Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0089s] [ 76%] 2025-07-17T09:17:08.5169140Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 76%] 2025-07-17T09:17:08.5169568Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0083s] [ 76%] 2025-07-17T09:17:08.5170003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 76%] 2025-07-17T09:17:08.5170433Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0069s] [ 76%] 2025-07-17T09:17:08.5170875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 76%] 2025-07-17T09:17:08.5171312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 76%] 2025-07-17T09:17:08.5171807Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 76%] 2025-07-17T09:17:08.5172286Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 76%] 2025-07-17T09:17:08.5172720Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 76%] 2025-07-17T09:17:08.5173275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 76%] 2025-07-17T09:17:08.5173709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 76%] 2025-07-17T09:17:08.5174155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 76%] 2025-07-17T09:17:08.5174596Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 76%] 2025-07-17T09:17:08.5175037Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0046s] [ 76%] 2025-07-17T09:17:08.5175476Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 76%] 2025-07-17T09:17:08.5175902Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0038s] [ 76%] 2025-07-17T09:17:08.5176338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 76%] 2025-07-17T09:17:08.5176769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 76%] 2025-07-17T09:17:08.5177209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 76%] 2025-07-17T09:17:08.5177636Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 76%] 2025-07-17T09:17:08.5178067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 76%] 2025-07-17T09:17:08.5178492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 76%] 2025-07-17T09:17:08.5178997Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 76%] 2025-07-17T09:17:08.5179489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 76%] 2025-07-17T09:17:08.5179923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 76%] 2025-07-17T09:17:08.5180467Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0049s] [ 76%] 2025-07-17T09:17:08.5180901Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 76%] 2025-07-17T09:17:08.5181336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0047s] [ 76%] 2025-07-17T09:17:08.5181764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 76%] 2025-07-17T09:17:08.5182203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0056s] [ 76%] 2025-07-17T09:17:08.5182643Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 76%] 2025-07-17T09:17:08.5183069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0051s] [ 76%] 2025-07-17T09:17:08.5183501Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 76%] 2025-07-17T09:17:08.5183927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 76%] 2025-07-17T09:17:08.5184370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 76%] 2025-07-17T09:17:08.5184807Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 76%] 2025-07-17T09:17:08.5185247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 76%] 2025-07-17T09:17:08.5185681Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0050s] [ 76%] 2025-07-17T09:17:08.5186197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 76%] 2025-07-17T09:17:08.5186677Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0038s] [ 76%] 2025-07-17T09:17:08.5187105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 76%] 2025-07-17T09:17:08.5187650Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 76%] 2025-07-17T09:17:08.5188089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 76%] 2025-07-17T09:17:08.5188524Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0047s] [ 76%] 2025-07-17T09:17:08.5188948Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 76%] 2025-07-17T09:17:08.5189381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 76%] 2025-07-17T09:17:08.5189812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 76%] 2025-07-17T09:17:08.5190240Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0058s] [ 77%] 2025-07-17T09:17:08.5190675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 77%] 2025-07-17T09:17:08.5191100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0058s] [ 77%] 2025-07-17T09:17:08.5191534Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 77%] 2025-07-17T09:17:08.5191963Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0046s] [ 77%] 2025-07-17T09:17:08.5192395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5192826Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0089s] [ 77%] 2025-07-17T09:17:08.5193325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 77%] 2025-07-17T09:17:08.5193806Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0079s] [ 77%] 2025-07-17T09:17:08.5194352Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 77%] 2025-07-17T09:17:08.5194936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0062s] [ 77%] 2025-07-17T09:17:08.5195378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 77%] 2025-07-17T09:17:08.5195818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0100s] [ 77%] 2025-07-17T09:17:08.5196257Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 77%] 2025-07-17T09:17:08.5196693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0097s] [ 77%] 2025-07-17T09:17:08.5197131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 77%] 2025-07-17T09:17:08.5197566Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0071s] [ 77%] 2025-07-17T09:17:08.5198002Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5198432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0080s] [ 77%] 2025-07-17T09:17:08.5198868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 77%] 2025-07-17T09:17:08.5199291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0074s] [ 77%] 2025-07-17T09:17:08.5199727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 77%] 2025-07-17T09:17:08.5200158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0066s] [ 77%] 2025-07-17T09:17:08.5200660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 77%] 2025-07-17T09:17:08.5201155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0080s] [ 77%] 2025-07-17T09:17:08.5201587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 77%] 2025-07-17T09:17:08.5202149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0077s] [ 77%] 2025-07-17T09:17:08.5202582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 77%] 2025-07-17T09:17:08.5203014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0073s] [ 77%] 2025-07-17T09:17:08.5203441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 77%] 2025-07-17T09:17:08.5203882Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 77%] 2025-07-17T09:17:08.5204323Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 77%] 2025-07-17T09:17:08.5204754Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 77%] 2025-07-17T09:17:08.5205188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 77%] 2025-07-17T09:17:08.5205616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 77%] 2025-07-17T09:17:08.5206049Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 77%] 2025-07-17T09:17:08.5206483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0048s] [ 77%] 2025-07-17T09:17:08.5206924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 77%] 2025-07-17T09:17:08.5207353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0048s] [ 77%] 2025-07-17T09:17:08.5207849Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 77%] 2025-07-17T09:17:08.5208334Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5208823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 77%] 2025-07-17T09:17:08.5210227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 77%] 2025-07-17T09:17:08.5210660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 77%] 2025-07-17T09:17:08.5211091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 77%] 2025-07-17T09:17:08.5211520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 77%] 2025-07-17T09:17:08.5211955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 77%] 2025-07-17T09:17:08.5212389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 77%] 2025-07-17T09:17:08.5212818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 77%] 2025-07-17T09:17:08.5213260Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 77%] 2025-07-17T09:17:08.5213684Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 77%] 2025-07-17T09:17:08.5214114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 77%] 2025-07-17T09:17:08.5214538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0046s] [ 77%] 2025-07-17T09:17:08.5214972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 77%] 2025-07-17T09:17:08.5215400Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 77%] 2025-07-17T09:17:08.5215902Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 77%] 2025-07-17T09:17:08.5216380Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 77%] 2025-07-17T09:17:08.5216857Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 77%] 2025-07-17T09:17:08.5217340Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 77%] 2025-07-17T09:17:08.5217769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 77%] 2025-07-17T09:17:08.5218210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 77%] 2025-07-17T09:17:08.5218645Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5219076Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 77%] 2025-07-17T09:17:08.5219510Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5219936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0032s] [ 77%] 2025-07-17T09:17:08.5220368Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5220792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 77%] 2025-07-17T09:17:08.5221230Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 77%] 2025-07-17T09:17:08.5221654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 77%] 2025-07-17T09:17:08.5222087Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 77%] 2025-07-17T09:17:08.5222510Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 77%] 2025-07-17T09:17:08.5222997Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 77%] 2025-07-17T09:17:08.5223480Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0050s] [ 77%] 2025-07-17T09:17:08.5223970Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 77%] 2025-07-17T09:17:08.5224451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0048s] [ 77%] 2025-07-17T09:17:08.5224875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 77%] 2025-07-17T09:17:08.5225303Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5225728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5226162Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0057s] [ 77%] 2025-07-17T09:17:08.5226597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 77%] 2025-07-17T09:17:08.5227026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0052s] [ 77%] 2025-07-17T09:17:08.5227459Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 77%] 2025-07-17T09:17:08.5227880Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5228312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 77%] 2025-07-17T09:17:08.5228744Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 77%] 2025-07-17T09:17:08.5229181Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5229604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0040s] [ 77%] 2025-07-17T09:17:08.5230089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 77%] 2025-07-17T09:17:08.5230564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5231064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 77%] 2025-07-17T09:17:08.5231544Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 77%] 2025-07-17T09:17:08.5231975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 77%] 2025-07-17T09:17:08.5232405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0048s] [ 77%] 2025-07-17T09:17:08.5232830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 77%] 2025-07-17T09:17:08.5233258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 77%] 2025-07-17T09:17:08.5233681Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 77%] 2025-07-17T09:17:08.5234189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0061s] [ 77%] 2025-07-17T09:17:08.5234628Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 77%] 2025-07-17T09:17:08.5235055Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0061s] [ 77%] 2025-07-17T09:17:08.5235484Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 77%] 2025-07-17T09:17:08.5235904Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 77%] 2025-07-17T09:17:08.5236337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 77%] 2025-07-17T09:17:08.5236764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 77%] 2025-07-17T09:17:08.5237285Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 77%] 2025-07-17T09:17:08.5237775Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 77%] 2025-07-17T09:17:08.5238266Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 77%] 2025-07-17T09:17:08.5238756Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 77%] 2025-07-17T09:17:08.5239188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0027s] [ 77%] 2025-07-17T09:17:08.5239627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 77%] 2025-07-17T09:17:08.5240062Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5240493Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0047s] [ 77%] 2025-07-17T09:17:08.5240923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 77%] 2025-07-17T09:17:08.5241354Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0041s] [ 77%] 2025-07-17T09:17:08.5241788Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 77%] 2025-07-17T09:17:08.5242211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0031s] [ 77%] 2025-07-17T09:17:08.5242653Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 77%] 2025-07-17T09:17:08.5243074Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 78%] 2025-07-17T09:17:08.5243509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 78%] 2025-07-17T09:17:08.5243930Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 78%] 2025-07-17T09:17:08.5244412Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 78%] 2025-07-17T09:17:08.5244897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5245380Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5245857Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5246283Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5246711Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0048s] [ 78%] 2025-07-17T09:17:08.5247137Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5247580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 78%] 2025-07-17T09:17:08.5248019Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 78%] 2025-07-17T09:17:08.5248460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 78%] 2025-07-17T09:17:08.5248901Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5249328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0032s] [ 78%] 2025-07-17T09:17:08.5249764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 78%] 2025-07-17T09:17:08.5250205Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5250654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5251084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5251582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 78%] 2025-07-17T09:17:08.5252063Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0045s] [ 78%] 2025-07-17T09:17:08.5252603Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 78%] 2025-07-17T09:17:08.5253042Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5253480Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5253916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5254347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5254780Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 78%] 2025-07-17T09:17:08.5255208Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 78%] 2025-07-17T09:17:08.5255645Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5256091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5256517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5256954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 78%] 2025-07-17T09:17:08.5257378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0038s] [ 78%] 2025-07-17T09:17:08.5257825Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 78%] 2025-07-17T09:17:08.5258329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5258835Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 78%] 2025-07-17T09:17:08.5259265Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 78%] 2025-07-17T09:17:08.5259798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 78%] 2025-07-17T09:17:08.5260238Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 78%] 2025-07-17T09:17:08.5260672Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 78%] 2025-07-17T09:17:08.5261115Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 78%] 2025-07-17T09:17:08.5261555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 78%] 2025-07-17T09:17:08.5261993Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0038s] [ 78%] 2025-07-17T09:17:08.5262425Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5262862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 78%] 2025-07-17T09:17:08.5263303Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5263729Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 78%] 2025-07-17T09:17:08.5264168Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5264591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 78%] 2025-07-17T09:17:08.5265025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 78%] 2025-07-17T09:17:08.5265506Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5265996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 78%] 2025-07-17T09:17:08.5266431Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 78%] 2025-07-17T09:17:08.5266976Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 78%] 2025-07-17T09:17:08.5267410Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 78%] 2025-07-17T09:17:08.5267840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5268267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 78%] 2025-07-17T09:17:08.5268700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5269135Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5269570Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 78%] 2025-07-17T09:17:08.5270000Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 78%] 2025-07-17T09:17:08.5270435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 78%] 2025-07-17T09:17:08.5270861Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0028s] [ 78%] 2025-07-17T09:17:08.5271297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 78%] 2025-07-17T09:17:08.5271729Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5272171Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5272650Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5273139Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5273573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 78%] 2025-07-17T09:17:08.5274184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 78%] 2025-07-17T09:17:08.5274625Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0031s] [ 78%] 2025-07-17T09:17:08.5275056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 78%] 2025-07-17T09:17:08.5275487Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5275916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 78%] 2025-07-17T09:17:08.5276353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 78%] 2025-07-17T09:17:08.5276780Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 78%] 2025-07-17T09:17:08.5277214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5277654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5278078Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5278511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5278931Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5279364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5279861Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5280360Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 78%] 2025-07-17T09:17:08.5280790Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 78%] 2025-07-17T09:17:08.5281339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 78%] 2025-07-17T09:17:08.5281774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 78%] 2025-07-17T09:17:08.5282206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 78%] 2025-07-17T09:17:08.5282648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5283090Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 78%] 2025-07-17T09:17:08.5283523Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5283957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 78%] 2025-07-17T09:17:08.5284390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5284844Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 78%] 2025-07-17T09:17:08.5285274Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 78%] 2025-07-17T09:17:08.5285712Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 78%] 2025-07-17T09:17:08.5286134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 78%] 2025-07-17T09:17:08.5286569Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 78%] 2025-07-17T09:17:08.5287058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 78%] 2025-07-17T09:17:08.5287540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 78%] 2025-07-17T09:17:08.5287972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5288515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5288955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5289384Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5289814Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0039s] [ 78%] 2025-07-17T09:17:08.5290244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 78%] 2025-07-17T09:17:08.5290679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0051s] [ 78%] 2025-07-17T09:17:08.5291114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 78%] 2025-07-17T09:17:08.5291538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 78%] 2025-07-17T09:17:08.5291974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 78%] 2025-07-17T09:17:08.5292402Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 78%] 2025-07-17T09:17:08.5292836Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 78%] 2025-07-17T09:17:08.5293265Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 78%] 2025-07-17T09:17:08.5293708Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 78%] 2025-07-17T09:17:08.5294190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0048s] [ 78%] 2025-07-17T09:17:08.5294678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 78%] 2025-07-17T09:17:08.5295112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0042s] [ 79%] 2025-07-17T09:17:08.5295640Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 79%] 2025-07-17T09:17:08.5296073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 79%] 2025-07-17T09:17:08.5296507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 79%] 2025-07-17T09:17:08.5296938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 79%] 2025-07-17T09:17:08.5297365Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 79%] 2025-07-17T09:17:08.5297794Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0044s] [ 79%] 2025-07-17T09:17:08.5298220Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 79%] 2025-07-17T09:17:08.5298651Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 79%] 2025-07-17T09:17:08.5299088Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 79%] 2025-07-17T09:17:08.5299509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0043s] [ 79%] 2025-07-17T09:17:08.5299952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 79%] 2025-07-17T09:17:08.5300377Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 79%] 2025-07-17T09:17:08.5300808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 79%] 2025-07-17T09:17:08.5301284Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 79%] 2025-07-17T09:17:08.5301769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0028s] [ 79%] 2025-07-17T09:17:08.5302195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0035s] [ 79%] 2025-07-17T09:17:08.5302723Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 79%] 2025-07-17T09:17:08.5303155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 79%] 2025-07-17T09:17:08.5303585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 79%] 2025-07-17T09:17:08.5304021Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 79%] 2025-07-17T09:17:08.5304457Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 79%] 2025-07-17T09:17:08.5304886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0042s] [ 79%] 2025-07-17T09:17:08.5305310Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 79%] 2025-07-17T09:17:08.5305732Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0039s] [ 79%] 2025-07-17T09:17:08.5306164Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 79%] 2025-07-17T09:17:08.5306585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 79%] 2025-07-17T09:17:08.5307020Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 79%] 2025-07-17T09:17:08.5307438Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0039s] [ 79%] 2025-07-17T09:17:08.5307869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 79%] 2025-07-17T09:17:08.5308340Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 79%] 2025-07-17T09:17:08.5308815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 79%] 2025-07-17T09:17:08.5309247Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 79%] 2025-07-17T09:17:08.5309800Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 79%] 2025-07-17T09:17:08.5310224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 79%] 2025-07-17T09:17:08.5310648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 79%] 2025-07-17T09:17:08.5311137Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 79%] 2025-07-17T09:17:08.5311742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 79%] 2025-07-17T09:17:08.5312220Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 79%] 2025-07-17T09:17:08.5318087Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 79%] 2025-07-17T09:17:08.5318540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0040s] [ 79%] 2025-07-17T09:17:08.5318976Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 79%] 2025-07-17T09:17:08.5319403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 79%] 2025-07-17T09:17:08.5319830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 79%] 2025-07-17T09:17:08.5320264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0061s] [ 79%] 2025-07-17T09:17:08.5320699Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 79%] 2025-07-17T09:17:08.5321724Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0055s] [ 79%] 2025-07-17T09:17:08.5322211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 79%] 2025-07-17T09:17:08.5322637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 79%] 2025-07-17T09:17:08.5323195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 79%] 2025-07-17T09:17:08.5323627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 79%] 2025-07-17T09:17:08.5324053Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 79%] 2025-07-17T09:17:08.5324480Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 79%] 2025-07-17T09:17:08.5324908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 79%] 2025-07-17T09:17:08.5325328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 79%] 2025-07-17T09:17:08.5325760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 79%] 2025-07-17T09:17:08.5326185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 79%] 2025-07-17T09:17:08.5326618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 79%] 2025-07-17T09:17:08.5327033Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 79%] 2025-07-17T09:17:08.5327462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 79%] 2025-07-17T09:17:08.5327880Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0039s] [ 79%] 2025-07-17T09:17:08.5328305Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 79%] 2025-07-17T09:17:08.5328792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0064s] [ 79%] 2025-07-17T09:17:08.5329278Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 79%] 2025-07-17T09:17:08.5329702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0054s] [ 79%] 2025-07-17T09:17:08.5330225Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 79%] 2025-07-17T09:17:08.5330655Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0039s] [ 79%] 2025-07-17T09:17:08.5331081Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 79%] 2025-07-17T09:17:08.5331509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 79%] 2025-07-17T09:17:08.5331948Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 79%] 2025-07-17T09:17:08.5332365Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0041s] [ 79%] 2025-07-17T09:17:08.5332793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 79%] 2025-07-17T09:17:08.5333213Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0050s] [ 79%] 2025-07-17T09:17:08.5333641Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 79%] 2025-07-17T09:17:08.5334060Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 79%] 2025-07-17T09:17:08.5334492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 79%] 2025-07-17T09:17:08.5334912Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 79%] 2025-07-17T09:17:08.5335332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 79%] 2025-07-17T09:17:08.5335750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 79%] 2025-07-17T09:17:08.5336241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 79%] 2025-07-17T09:17:08.5336719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 79%] 2025-07-17T09:17:08.5337253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 79%] 2025-07-17T09:17:08.5337683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0043s] [ 79%] 2025-07-17T09:17:08.5338105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 79%] 2025-07-17T09:17:08.5338532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 79%] 2025-07-17T09:17:08.5338959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 79%] 2025-07-17T09:17:08.5339400Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0495s] [ 79%] 2025-07-17T09:17:08.5339850Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0062s] [ 79%] 2025-07-17T09:17:08.5340281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0069s] [ 79%] 2025-07-17T09:17:08.5340724Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0060s] [ 79%] 2025-07-17T09:17:08.5341154Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0061s] [ 79%] 2025-07-17T09:17:08.5341595Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0057s] [ 79%] 2025-07-17T09:17:08.5342169Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0030s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 79%] 2025-07-17T09:17:08.5342734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 79%] 2025-07-17T09:17:08.5343373Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 79%] 2025-07-17T09:17:08.5343991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0033s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 79%] 2025-07-17T09:17:08.5344655Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 79%] 2025-07-17T09:17:08.5345214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 79%] 2025-07-17T09:17:08.5345661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 79%] 2025-07-17T09:17:08.5346107Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0056s] [ 79%] 2025-07-17T09:17:08.5346535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0056s] [ 79%] 2025-07-17T09:17:08.5346970Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0061s] [ 79%] 2025-07-17T09:17:08.5347397Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0062s] [ 79%] 2025-07-17T09:17:08.5347837Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0058s] [ 79%] 2025-07-17T09:17:08.5348405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0030s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 79%] 2025-07-17T09:17:08.5348972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0047s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 79%] 2025-07-17T09:17:08.5349526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 79%] 2025-07-17T09:17:08.5350076Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 79%] 2025-07-17T09:17:08.5350689Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 79%] 2025-07-17T09:17:08.5351286Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 79%] 2025-07-17T09:17:08.5351824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0118s] [ 79%] 2025-07-17T09:17:08.5352276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 79%] 2025-07-17T09:17:08.5352705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0101s] [ 79%] 2025-07-17T09:17:08.5353140Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 79%] 2025-07-17T09:17:08.5353573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0072s] [ 79%] 2025-07-17T09:17:08.5354097Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 79%] 2025-07-17T09:17:08.5354651Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5355209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5355763Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5356313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0021s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5356865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5357414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5357918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0120s] [ 80%] 2025-07-17T09:17:08.5358419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 80%] 2025-07-17T09:17:08.5359002Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0112s] [ 80%] 2025-07-17T09:17:08.5359445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 80%] 2025-07-17T09:17:08.5359876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0087s] [ 80%] 2025-07-17T09:17:08.5360311Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 80%] 2025-07-17T09:17:08.5360865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5361424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5361975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5362523Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5363075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5363620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5364059Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 80%] 2025-07-17T09:17:08.5364495Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 80%] 2025-07-17T09:17:08.5364981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 80%] 2025-07-17T09:17:08.5365475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 80%] 2025-07-17T09:17:08.5365905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 80%] 2025-07-17T09:17:08.5366449Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 80%] 2025-07-17T09:17:08.5367008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5367567Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5368117Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5368672Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5369219Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5369768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5370204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 80%] 2025-07-17T09:17:08.5370637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 80%] 2025-07-17T09:17:08.5371069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 80%] 2025-07-17T09:17:08.5371511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 80%] 2025-07-17T09:17:08.5371989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 80%] 2025-07-17T09:17:08.5372471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 80%] 2025-07-17T09:17:08.5373018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5373683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5374249Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5374798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5375352Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5375901Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5376348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0459s] [ 80%] 2025-07-17T09:17:08.5376787Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0061s] [ 80%] 2025-07-17T09:17:08.5377223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0060s] [ 80%] 2025-07-17T09:17:08.5377652Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0060s] [ 80%] 2025-07-17T09:17:08.5378087Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0058s] [ 80%] 2025-07-17T09:17:08.5378531Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0074s] [ 80%] 2025-07-17T09:17:08.5379095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0029s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5379716Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5380317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5380985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0028s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5381542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5382095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5382543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 80%] 2025-07-17T09:17:08.5382996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0057s] [ 80%] 2025-07-17T09:17:08.5383433Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0058s] [ 80%] 2025-07-17T09:17:08.5383863Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0059s] [ 80%] 2025-07-17T09:17:08.5384304Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0057s] [ 80%] 2025-07-17T09:17:08.5384743Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0043s] [ 80%] 2025-07-17T09:17:08.5385301Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0029s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5385861Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0069s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5386407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5387021Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5387618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5388291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5388743Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0050s] [ 80%] 2025-07-17T09:17:08.5389186Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 80%] 2025-07-17T09:17:08.5389621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 80%] 2025-07-17T09:17:08.5390058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 80%] 2025-07-17T09:17:08.5390491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 80%] 2025-07-17T09:17:08.5390924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 80%] 2025-07-17T09:17:08.5391491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5392051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5392604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5393163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5393768Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5394447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5394889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0054s] [ 80%] 2025-07-17T09:17:08.5395464Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 80%] 2025-07-17T09:17:08.5395906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0045s] [ 80%] 2025-07-17T09:17:08.5396335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 80%] 2025-07-17T09:17:08.5396767Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0044s] [ 80%] 2025-07-17T09:17:08.5397200Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 80%] 2025-07-17T09:17:08.5397756Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5398309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5398863Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5399413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5399959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5400512Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5400948Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0057s] [ 80%] 2025-07-17T09:17:08.5401454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 80%] 2025-07-17T09:17:08.5401947Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0052s] [ 80%] 2025-07-17T09:17:08.5402490Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 80%] 2025-07-17T09:17:08.5402924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0045s] [ 80%] 2025-07-17T09:17:08.5403352Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 80%] 2025-07-17T09:17:08.5403909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5404466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5405014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5405562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5406108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5406659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5407087Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 80%] 2025-07-17T09:17:08.5407526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 80%] 2025-07-17T09:17:08.5407957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 80%] 2025-07-17T09:17:08.5408446Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 80%] 2025-07-17T09:17:08.5408932Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0047s] [ 80%] 2025-07-17T09:17:08.5409357Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 80%] 2025-07-17T09:17:08.5410023Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5410591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5411145Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5411695Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5412246Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5412794Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 80%] 2025-07-17T09:17:08.5413228Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 80%] 2025-07-17T09:17:08.5413671Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 80%] 2025-07-17T09:17:08.5414103Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 81%] 2025-07-17T09:17:08.5414532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 81%] 2025-07-17T09:17:08.5414966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 81%] 2025-07-17T09:17:08.5415447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 81%] 2025-07-17T09:17:08.5416069Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5416620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5417281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5417837Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5418385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5418936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5419366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 81%] 2025-07-17T09:17:08.5419801Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 81%] 2025-07-17T09:17:08.5420227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 81%] 2025-07-17T09:17:08.5420658Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 81%] 2025-07-17T09:17:08.5421084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 81%] 2025-07-17T09:17:08.5421513Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 81%] 2025-07-17T09:17:08.5422066Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5422671Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5423270Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5423808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5424461Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5425010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5425445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 81%] 2025-07-17T09:17:08.5425885Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 81%] 2025-07-17T09:17:08.5426310Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 81%] 2025-07-17T09:17:08.5426759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0043s] [ 81%] 2025-07-17T09:17:08.5427188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0039s] [ 81%] 2025-07-17T09:17:08.5427623Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 81%] 2025-07-17T09:17:08.5428186Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5428739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5429294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5429924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5430530Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5431135Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5431755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 81%] 2025-07-17T09:17:08.5432198Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 81%] 2025-07-17T09:17:08.5432622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 81%] 2025-07-17T09:17:08.5433051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 81%] 2025-07-17T09:17:08.5433475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 81%] 2025-07-17T09:17:08.5433987Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 81%] 2025-07-17T09:17:08.5434548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5435102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5435651Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5436197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5436745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5437365Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5437865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0371s] [ 81%] 2025-07-17T09:17:08.5438314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0069s] [ 81%] 2025-07-17T09:17:08.5438871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0067s] [ 81%] 2025-07-17T09:17:08.5439318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0069s] [ 81%] 2025-07-17T09:17:08.5439748Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0073s] [ 81%] 2025-07-17T09:17:08.5440195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0073s] [ 81%] 2025-07-17T09:17:08.5440762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0034s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5441321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5441879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5442436Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0029s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5442988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5443547Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5443988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 81%] 2025-07-17T09:17:08.5444429Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0069s] [ 81%] 2025-07-17T09:17:08.5444932Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0069s] [ 81%] 2025-07-17T09:17:08.5445420Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0069s] [ 81%] 2025-07-17T09:17:08.5445951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0074s] [ 81%] 2025-07-17T09:17:08.5446390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0072s] [ 81%] 2025-07-17T09:17:08.5446946Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0030s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5447506Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0046s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5448061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5448611Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5449164Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5449715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5450155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0060s] [ 81%] 2025-07-17T09:17:08.5450600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 81%] 2025-07-17T09:17:08.5451026Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0050s] [ 81%] 2025-07-17T09:17:08.5451465Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 81%] 2025-07-17T09:17:08.5451941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0055s] [ 81%] 2025-07-17T09:17:08.5452427Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 81%] 2025-07-17T09:17:08.5453036Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5453648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5454202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5454753Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5455307Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5455855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5456295Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 81%] 2025-07-17T09:17:08.5456738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 81%] 2025-07-17T09:17:08.5457165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0043s] [ 81%] 2025-07-17T09:17:08.5457597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 81%] 2025-07-17T09:17:08.5458020Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0052s] [ 81%] 2025-07-17T09:17:08.5458455Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 81%] 2025-07-17T09:17:08.5459068Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5459680Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5460291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5460891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5461438Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5461982Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5462425Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 81%] 2025-07-17T09:17:08.5462862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 81%] 2025-07-17T09:17:08.5463292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0039s] [ 81%] 2025-07-17T09:17:08.5463726Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 81%] 2025-07-17T09:17:08.5464163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0046s] [ 81%] 2025-07-17T09:17:08.5464602Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 81%] 2025-07-17T09:17:08.5465157Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5465716Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5466315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5466918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5467518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5468141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5468580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 81%] 2025-07-17T09:17:08.5469018Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 81%] 2025-07-17T09:17:08.5469452Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 81%] 2025-07-17T09:17:08.5469890Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 81%] 2025-07-17T09:17:08.5470316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0041s] [ 81%] 2025-07-17T09:17:08.5470748Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0041s] [ 81%] 2025-07-17T09:17:08.5471303Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5471866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5472417Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5472970Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 81%] 2025-07-17T09:17:08.5473597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5474264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5474784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0506s] [ 82%] 2025-07-17T09:17:08.5475286Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0067s] [ 82%] 2025-07-17T09:17:08.5475726Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0075s] [ 82%] 2025-07-17T09:17:08.5476167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0065s] [ 82%] 2025-07-17T09:17:08.5476605Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0071s] [ 82%] 2025-07-17T09:17:08.5477041Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0071s] [ 82%] 2025-07-17T09:17:08.5477609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0033s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5478168Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5478728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5479279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0029s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5479828Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5480384Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5480880Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0062s] [ 82%] 2025-07-17T09:17:08.5481381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0065s] [ 82%] 2025-07-17T09:17:08.5481810Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0070s] [ 82%] 2025-07-17T09:17:08.5482347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0064s] [ 82%] 2025-07-17T09:17:08.5482784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0068s] [ 82%] 2025-07-17T09:17:08.5483215Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0054s] [ 82%] 2025-07-17T09:17:08.5483771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0029s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5484324Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0040s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5484876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5485423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5485968Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5486516Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5486950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0048s] [ 82%] 2025-07-17T09:17:08.5487393Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 82%] 2025-07-17T09:17:08.5487887Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 82%] 2025-07-17T09:17:08.5488367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 82%] 2025-07-17T09:17:08.5488794Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 82%] 2025-07-17T09:17:08.5489330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 82%] 2025-07-17T09:17:08.5489891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5490443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5490995Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5491541Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5492089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5492637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5493063Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 82%] 2025-07-17T09:17:08.5493498Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 82%] 2025-07-17T09:17:08.5493921Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 82%] 2025-07-17T09:17:08.5494345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 82%] 2025-07-17T09:17:08.5494770Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 82%] 2025-07-17T09:17:08.5495250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 82%] 2025-07-17T09:17:08.5495846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5496471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5497071Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5497612Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5498155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5498703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5499132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0071s] [ 82%] 2025-07-17T09:17:08.5499565Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 82%] 2025-07-17T09:17:08.5499988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0054s] [ 82%] 2025-07-17T09:17:08.5500419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 82%] 2025-07-17T09:17:08.5500837Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0044s] [ 82%] 2025-07-17T09:17:08.5501260Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 82%] 2025-07-17T09:17:08.5501813Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5502419Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5503017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5503607Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5504199Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5504748Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5505173Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 82%] 2025-07-17T09:17:08.5505610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 82%] 2025-07-17T09:17:08.5506033Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0047s] [ 82%] 2025-07-17T09:17:08.5506463Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 82%] 2025-07-17T09:17:08.5506880Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0041s] [ 82%] 2025-07-17T09:17:08.5507312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 82%] 2025-07-17T09:17:08.5507864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5508413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5508963Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5509553Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5510148Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5510738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5511226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 82%] 2025-07-17T09:17:08.5511663Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 82%] 2025-07-17T09:17:08.5512085Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 82%] 2025-07-17T09:17:08.5512514Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 82%] 2025-07-17T09:17:08.5512933Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 82%] 2025-07-17T09:17:08.5513361Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 82%] 2025-07-17T09:17:08.5513980Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5514553Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5515096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5515640Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5516188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5516793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5517281Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 82%] 2025-07-17T09:17:08.5517720Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 82%] 2025-07-17T09:17:08.5518267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 82%] 2025-07-17T09:17:08.5518700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 82%] 2025-07-17T09:17:08.5519126Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 82%] 2025-07-17T09:17:08.5519556Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 82%] 2025-07-17T09:17:08.5520109Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5520658Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5521201Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5521745Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5522288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5522826Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5523262Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 82%] 2025-07-17T09:17:08.5523696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 82%] 2025-07-17T09:17:08.5524176Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0039s] [ 82%] 2025-07-17T09:17:08.5524667Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 82%] 2025-07-17T09:17:08.5525141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 82%] 2025-07-17T09:17:08.5525624Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 82%] 2025-07-17T09:17:08.5526173Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5526727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5527275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5527821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5528366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5528911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 82%] 2025-07-17T09:17:08.5529346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 82%] 2025-07-17T09:17:08.5529776Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 82%] 2025-07-17T09:17:08.5530203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0047s] [ 82%] 2025-07-17T09:17:08.5530629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 82%] 2025-07-17T09:17:08.5531109Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 82%] 2025-07-17T09:17:08.5531582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 82%] 2025-07-17T09:17:08.5532127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 83%] 2025-07-17T09:17:08.5532784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 83%] 2025-07-17T09:17:08.5533332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 83%] 2025-07-17T09:17:08.5533878Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 83%] 2025-07-17T09:17:08.5534423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 83%] 2025-07-17T09:17:08.5534963Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 83%] 2025-07-17T09:17:08.5535401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0061s] [ 83%] 2025-07-17T09:17:08.5535838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 83%] 2025-07-17T09:17:08.5536267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0054s] [ 83%] 2025-07-17T09:17:08.5536693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5537123Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0039s] [ 83%] 2025-07-17T09:17:08.5537552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 83%] 2025-07-17T09:17:08.5537987Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0050s] [ 83%] 2025-07-17T09:17:08.5538482Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5538960Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0048s] [ 83%] 2025-07-17T09:17:08.5539439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5539918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0042s] [ 83%] 2025-07-17T09:17:08.5540353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 83%] 2025-07-17T09:17:08.5540781Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 83%] 2025-07-17T09:17:08.5541217Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 83%] 2025-07-17T09:17:08.5541641Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0046s] [ 83%] 2025-07-17T09:17:08.5542064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 83%] 2025-07-17T09:17:08.5542489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 83%] 2025-07-17T09:17:08.5542915Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 83%] 2025-07-17T09:17:08.5543344Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 83%] 2025-07-17T09:17:08.5543774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 83%] 2025-07-17T09:17:08.5544198Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0049s] [ 83%] 2025-07-17T09:17:08.5544634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 83%] 2025-07-17T09:17:08.5545053Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0049s] [ 83%] 2025-07-17T09:17:08.5545533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5546005Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 83%] 2025-07-17T09:17:08.5546666Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 83%] 2025-07-17T09:17:08.5547141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 83%] 2025-07-17T09:17:08.5547572Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 83%] 2025-07-17T09:17:08.5547994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0028s] [ 83%] 2025-07-17T09:17:08.5548432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 83%] 2025-07-17T09:17:08.5548874Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 83%] 2025-07-17T09:17:08.5549307Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 83%] 2025-07-17T09:17:08.5549737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0040s] [ 83%] 2025-07-17T09:17:08.5550161Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 83%] 2025-07-17T09:17:08.5550588Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 83%] 2025-07-17T09:17:08.5551014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 83%] 2025-07-17T09:17:08.5551448Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 83%] 2025-07-17T09:17:08.5551876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 83%] 2025-07-17T09:17:08.5552302Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0043s] [ 83%] 2025-07-17T09:17:08.5552795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 83%] 2025-07-17T09:17:08.5553261Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5553741Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 83%] 2025-07-17T09:17:08.5554276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 83%] 2025-07-17T09:17:08.5554713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 83%] 2025-07-17T09:17:08.5555134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0048s] [ 83%] 2025-07-17T09:17:08.5555562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5555983Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0047s] [ 83%] 2025-07-17T09:17:08.5556413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 83%] 2025-07-17T09:17:08.5556840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0029s] [ 83%] 2025-07-17T09:17:08.5557268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 83%] 2025-07-17T09:17:08.5557690Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0029s] [ 83%] 2025-07-17T09:17:08.5558113Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 83%] 2025-07-17T09:17:08.5558536Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 83%] 2025-07-17T09:17:08.5558959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 83%] 2025-07-17T09:17:08.5559388Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5559894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 83%] 2025-07-17T09:17:08.5560376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 83%] 2025-07-17T09:17:08.5560868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 83%] 2025-07-17T09:17:08.5561345Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0045s] [ 83%] 2025-07-17T09:17:08.5561779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 83%] 2025-07-17T09:17:08.5562203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 83%] 2025-07-17T09:17:08.5562632Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 83%] 2025-07-17T09:17:08.5563054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 83%] 2025-07-17T09:17:08.5563478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 83%] 2025-07-17T09:17:08.5563900Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 83%] 2025-07-17T09:17:08.5564321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 83%] 2025-07-17T09:17:08.5564750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0067s] [ 83%] 2025-07-17T09:17:08.5565178Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 83%] 2025-07-17T09:17:08.5565601Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0063s] [ 83%] 2025-07-17T09:17:08.5566023Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5566443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0047s] [ 83%] 2025-07-17T09:17:08.5566922Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 83%] 2025-07-17T09:17:08.5567403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 83%] 2025-07-17T09:17:08.5567839Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 83%] 2025-07-17T09:17:08.5568380Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5568814Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 83%] 2025-07-17T09:17:08.5569236Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 83%] 2025-07-17T09:17:08.5569662Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 83%] 2025-07-17T09:17:08.5570096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 83%] 2025-07-17T09:17:08.5570537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 83%] 2025-07-17T09:17:08.5570968Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0041s] [ 83%] 2025-07-17T09:17:08.5571396Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5571824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 83%] 2025-07-17T09:17:08.5572249Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 83%] 2025-07-17T09:17:08.5572675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 83%] 2025-07-17T09:17:08.5573100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 83%] 2025-07-17T09:17:08.5573521Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0045s] [ 83%] 2025-07-17T09:17:08.5574028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 83%] 2025-07-17T09:17:08.5574497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0043s] [ 83%] 2025-07-17T09:17:08.5574926Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 83%] 2025-07-17T09:17:08.5575481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 83%] 2025-07-17T09:17:08.5575927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5576346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0041s] [ 83%] 2025-07-17T09:17:08.5576772Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5577194Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0041s] [ 83%] 2025-07-17T09:17:08.5577626Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 83%] 2025-07-17T09:17:08.5578055Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 83%] 2025-07-17T09:17:08.5578489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 83%] 2025-07-17T09:17:08.5578917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0034s] [ 83%] 2025-07-17T09:17:08.5579342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 83%] 2025-07-17T09:17:08.5579766Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5580189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 83%] 2025-07-17T09:17:08.5580621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 83%] 2025-07-17T09:17:08.5581108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 83%] 2025-07-17T09:17:08.5581581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 83%] 2025-07-17T09:17:08.5582010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 83%] 2025-07-17T09:17:08.5582536Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0041s] [ 83%] 2025-07-17T09:17:08.5582972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 83%] 2025-07-17T09:17:08.5583396Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 83%] 2025-07-17T09:17:08.5583824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 84%] 2025-07-17T09:17:08.5584243Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0044s] [ 84%] 2025-07-17T09:17:08.5584670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 84%] 2025-07-17T09:17:08.5585089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5585511Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5585941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 84%] 2025-07-17T09:17:08.5586373Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5586797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0041s] [ 84%] 2025-07-17T09:17:08.5587221Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 84%] 2025-07-17T09:17:08.5587642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5588125Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5588596Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 84%] 2025-07-17T09:17:08.5589027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 84%] 2025-07-17T09:17:08.5589558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 84%] 2025-07-17T09:17:08.5589988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5590408Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 84%] 2025-07-17T09:17:08.5590830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0027s] [ 84%] 2025-07-17T09:17:08.5591258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 84%] 2025-07-17T09:17:08.5591694Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 84%] 2025-07-17T09:17:08.5592114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0046s] [ 84%] 2025-07-17T09:17:08.5592538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 84%] 2025-07-17T09:17:08.5592958Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0041s] [ 84%] 2025-07-17T09:17:08.5593382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 84%] 2025-07-17T09:17:08.5593803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 84%] 2025-07-17T09:17:08.5594297Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 84%] 2025-07-17T09:17:08.5594730Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 84%] 2025-07-17T09:17:08.5595214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 84%] 2025-07-17T09:17:08.5595689Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0047s] [ 84%] 2025-07-17T09:17:08.5596112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 84%] 2025-07-17T09:17:08.5596659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0052s] [ 84%] 2025-07-17T09:17:08.5597095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 84%] 2025-07-17T09:17:08.5597512Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0050s] [ 84%] 2025-07-17T09:17:08.5597941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 84%] 2025-07-17T09:17:08.5598361Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5598791Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5599213Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 84%] 2025-07-17T09:17:08.5599638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5600057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5600478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5600898Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0043s] [ 84%] 2025-07-17T09:17:08.5601320Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 84%] 2025-07-17T09:17:08.5601754Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 84%] 2025-07-17T09:17:08.5602234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 84%] 2025-07-17T09:17:08.5602707Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0033s] [ 84%] 2025-07-17T09:17:08.5603133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 84%] 2025-07-17T09:17:08.5603678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 84%] 2025-07-17T09:17:08.5604110Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 84%] 2025-07-17T09:17:08.5604530Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0053s] [ 84%] 2025-07-17T09:17:08.5604957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 84%] 2025-07-17T09:17:08.5605373Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0052s] [ 84%] 2025-07-17T09:17:08.5605792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 84%] 2025-07-17T09:17:08.5606216Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0042s] [ 84%] 2025-07-17T09:17:08.5606637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5607062Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0055s] [ 84%] 2025-07-17T09:17:08.5607491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 84%] 2025-07-17T09:17:08.5607913Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0055s] [ 84%] 2025-07-17T09:17:08.5608331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 84%] 2025-07-17T09:17:08.5608755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5609237Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5609724Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 84%] 2025-07-17T09:17:08.5610155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5610686Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5611118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5611535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 84%] 2025-07-17T09:17:08.5611959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 84%] 2025-07-17T09:17:08.5612387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 84%] 2025-07-17T09:17:08.5612821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 84%] 2025-07-17T09:17:08.5613245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0033s] [ 84%] 2025-07-17T09:17:08.5613670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5614093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 84%] 2025-07-17T09:17:08.5614517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 84%] 2025-07-17T09:17:08.5614942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 84%] 2025-07-17T09:17:08.5615366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 84%] 2025-07-17T09:17:08.5615794Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0039s] [ 84%] 2025-07-17T09:17:08.5616269Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 84%] 2025-07-17T09:17:08.5616742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5617167Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5617696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 84%] 2025-07-17T09:17:08.5618133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5618551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5618975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5619395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0042s] [ 84%] 2025-07-17T09:17:08.5619820Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5620250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 84%] 2025-07-17T09:17:08.5620693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 84%] 2025-07-17T09:17:08.5621123Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 84%] 2025-07-17T09:17:08.5621551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 84%] 2025-07-17T09:17:08.5621980Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 84%] 2025-07-17T09:17:08.5622406Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 84%] 2025-07-17T09:17:08.5622840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 84%] 2025-07-17T09:17:08.5623331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 84%] 2025-07-17T09:17:08.5623808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 84%] 2025-07-17T09:17:08.5624251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 84%] 2025-07-17T09:17:08.5624780Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 84%] 2025-07-17T09:17:08.5625218Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 84%] 2025-07-17T09:17:08.5625645Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 84%] 2025-07-17T09:17:08.5626077Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 84%] 2025-07-17T09:17:08.5626506Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0034s] [ 84%] 2025-07-17T09:17:08.5626932Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 84%] 2025-07-17T09:17:08.5627351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0032s] [ 84%] 2025-07-17T09:17:08.5627778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 84%] 2025-07-17T09:17:08.5628210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 84%] 2025-07-17T09:17:08.5628648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 84%] 2025-07-17T09:17:08.5629077Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0038s] [ 84%] 2025-07-17T09:17:08.5629500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 84%] 2025-07-17T09:17:08.5629929Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 84%] 2025-07-17T09:17:08.5630430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 84%] 2025-07-17T09:17:08.5630937Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 84%] 2025-07-17T09:17:08.5631372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 84%] 2025-07-17T09:17:08.5631893Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 84%] 2025-07-17T09:17:08.5632330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 84%] 2025-07-17T09:17:08.5632762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 84%] 2025-07-17T09:17:08.5633193Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 84%] 2025-07-17T09:17:08.5633628Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 84%] 2025-07-17T09:17:08.5634132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 84%] 2025-07-17T09:17:08.5634557Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 84%] 2025-07-17T09:17:08.5634991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5635418Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0043s] [ 85%] 2025-07-17T09:17:08.5635844Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5636275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 85%] 2025-07-17T09:17:08.5636702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 85%] 2025-07-17T09:17:08.5637127Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 85%] 2025-07-17T09:17:08.5637618Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 85%] 2025-07-17T09:17:08.5638102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 85%] 2025-07-17T09:17:08.5638534Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5639090Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 85%] 2025-07-17T09:17:08.5639531Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 85%] 2025-07-17T09:17:08.5639955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 85%] 2025-07-17T09:17:08.5640383Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 85%] 2025-07-17T09:17:08.5640805Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 85%] 2025-07-17T09:17:08.5641236Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 85%] 2025-07-17T09:17:08.5641660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5642095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5642527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5642951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5643377Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 85%] 2025-07-17T09:17:08.5643804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 85%] 2025-07-17T09:17:08.5644234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5644751Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5645223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5645648Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5646182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0041s] [ 85%] 2025-07-17T09:17:08.5646619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 85%] 2025-07-17T09:17:08.5647042Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 85%] 2025-07-17T09:17:08.5647474Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 85%] 2025-07-17T09:17:08.5647896Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 85%] 2025-07-17T09:17:08.5648325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 85%] 2025-07-17T09:17:08.5648740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5649166Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 85%] 2025-07-17T09:17:08.5649595Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5650025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5650445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5650867Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5651296Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 85%] 2025-07-17T09:17:08.5651721Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 85%] 2025-07-17T09:17:08.5652207Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0031s] [ 85%] 2025-07-17T09:17:08.5652685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5653230Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 85%] 2025-07-17T09:17:08.5653663Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5654085Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 85%] 2025-07-17T09:17:08.5654517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 85%] 2025-07-17T09:17:08.5654951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5655395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5655821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5656252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5656679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5657107Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5657535Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 85%] 2025-07-17T09:17:08.5657965Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 85%] 2025-07-17T09:17:08.5658387Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 85%] 2025-07-17T09:17:08.5658812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 85%] 2025-07-17T09:17:08.5659309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5659901Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 85%] 2025-07-17T09:17:08.5660441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0050s] [ 85%] 2025-07-17T09:17:08.5660877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 85%] 2025-07-17T09:17:08.5661299Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0044s] [ 85%] 2025-07-17T09:17:08.5661727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 85%] 2025-07-17T09:17:08.5662151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0043s] [ 85%] 2025-07-17T09:17:08.5662580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 85%] 2025-07-17T09:17:08.5663003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 85%] 2025-07-17T09:17:08.5663440Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5663866Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5664290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 85%] 2025-07-17T09:17:08.5664710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0028s] [ 85%] 2025-07-17T09:17:08.5665130Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 85%] 2025-07-17T09:17:08.5665559Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 85%] 2025-07-17T09:17:08.5665988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 85%] 2025-07-17T09:17:08.5666462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0043s] [ 85%] 2025-07-17T09:17:08.5666934Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 85%] 2025-07-17T09:17:08.5667464Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0032s] [ 85%] 2025-07-17T09:17:08.5667899Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 85%] 2025-07-17T09:17:08.5668323Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0031s] [ 85%] 2025-07-17T09:17:08.5668752Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5669171Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5669594Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5670012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 85%] 2025-07-17T09:17:08.5670441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 85%] 2025-07-17T09:17:08.5670864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 85%] 2025-07-17T09:17:08.5671295Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 85%] 2025-07-17T09:17:08.5671715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0040s] [ 85%] 2025-07-17T09:17:08.5672136Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 85%] 2025-07-17T09:17:08.5672558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0042s] [ 85%] 2025-07-17T09:17:08.5672978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 85%] 2025-07-17T09:17:08.5673457Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 85%] 2025-07-17T09:17:08.5674006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 85%] 2025-07-17T09:17:08.5674492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 85%] 2025-07-17T09:17:08.5674978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 85%] 2025-07-17T09:17:08.5675397Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 85%] 2025-07-17T09:17:08.5675822Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 85%] 2025-07-17T09:17:08.5676243Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 85%] 2025-07-17T09:17:08.5676678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 85%] 2025-07-17T09:17:08.5677097Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 85%] 2025-07-17T09:17:08.5677524Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 85%] 2025-07-17T09:17:08.5677938Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0038s] [ 85%] 2025-07-17T09:17:08.5678369Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 85%] 2025-07-17T09:17:08.5678790Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 85%] 2025-07-17T09:17:08.5679210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 85%] 2025-07-17T09:17:08.5679630Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 85%] 2025-07-17T09:17:08.5680051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 85%] 2025-07-17T09:17:08.5680533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0042s] [ 85%] 2025-07-17T09:17:08.5681017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 85%] 2025-07-17T09:17:08.5681441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 85%] 2025-07-17T09:17:08.5681989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 85%] 2025-07-17T09:17:08.5682417Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0042s] [ 85%] 2025-07-17T09:17:08.5682842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 85%] 2025-07-17T09:17:08.5683254Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0047s] [ 85%] 2025-07-17T09:17:08.5683679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 85%] 2025-07-17T09:17:08.5684100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0029s] [ 85%] 2025-07-17T09:17:08.5684533Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0028s] [ 85%] 2025-07-17T09:17:08.5684945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0028s] [ 85%] 2025-07-17T09:17:08.5685368Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 85%] 2025-07-17T09:17:08.5685789Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 85%] 2025-07-17T09:17:08.5686212Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 86%] 2025-07-17T09:17:08.5686638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 86%] 2025-07-17T09:17:08.5687065Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 86%] 2025-07-17T09:17:08.5687544Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 86%] 2025-07-17T09:17:08.5688012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 86%] 2025-07-17T09:17:08.5688428Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 86%] 2025-07-17T09:17:08.5688951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 86%] 2025-07-17T09:17:08.5689378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0031s] [ 86%] 2025-07-17T09:17:08.5689808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 86%] 2025-07-17T09:17:08.5690223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 86%] 2025-07-17T09:17:08.5690644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 86%] 2025-07-17T09:17:08.5691055Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0043s] [ 86%] 2025-07-17T09:17:08.5691474Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 86%] 2025-07-17T09:17:08.5691889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 86%] 2025-07-17T09:17:08.5692318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 86%] 2025-07-17T09:17:08.5692732Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 86%] 2025-07-17T09:17:08.5693156Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 86%] 2025-07-17T09:17:08.5693573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0049s] [ 86%] 2025-07-17T09:17:08.5693993Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 86%] 2025-07-17T09:17:08.5694416Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 86%] 2025-07-17T09:17:08.5694911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 86%] 2025-07-17T09:17:08.5695378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 86%] 2025-07-17T09:17:08.5695902Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 86%] 2025-07-17T09:17:08.5696325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 86%] 2025-07-17T09:17:08.5696753Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 86%] 2025-07-17T09:17:08.5697182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 86%] 2025-07-17T09:17:08.5697616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 86%] 2025-07-17T09:17:08.5698033Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0041s] [ 86%] 2025-07-17T09:17:08.5698465Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 86%] 2025-07-17T09:17:08.5698879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0041s] [ 86%] 2025-07-17T09:17:08.5699307Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 86%] 2025-07-17T09:17:08.5699727Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 86%] 2025-07-17T09:17:08.5700155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 86%] 2025-07-17T09:17:08.5700573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 86%] 2025-07-17T09:17:08.5700997Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 86%] 2025-07-17T09:17:08.5701412Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0044s] [ 86%] 2025-07-17T09:17:08.5701892Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 86%] 2025-07-17T09:17:08.5702367Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 86%] 2025-07-17T09:17:08.5702844Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 86%] 2025-07-17T09:17:08.5703316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0040s] [ 86%] 2025-07-17T09:17:08.5703736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 86%] 2025-07-17T09:17:08.5704157Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0048s] [ 86%] 2025-07-17T09:17:08.5704583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_1_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 86%] 2025-07-17T09:17:08.5705023Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0452s] [ 86%] 2025-07-17T09:17:08.5705470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0105s] [ 86%] 2025-07-17T09:17:08.5705901Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0103s] [ 86%] 2025-07-17T09:17:08.5706343Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0103s] [ 86%] 2025-07-17T09:17:08.5706772Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0115s] [ 86%] 2025-07-17T09:17:08.5707215Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0116s] [ 86%] 2025-07-17T09:17:08.5707657Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0125s] [ 86%] 2025-07-17T09:17:08.5708099Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0132s] [ 86%] 2025-07-17T09:17:08.5708538Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0132s] [ 86%] 2025-07-17T09:17:08.5709060Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0131s] [ 86%] 2025-07-17T09:17:08.5709551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0142s] [ 86%] 2025-07-17T09:17:08.5710126Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0147s] [ 86%] 2025-07-17T09:17:08.5710565Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0122s] [ 86%] 2025-07-17T09:17:08.5711003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0106s] [ 86%] 2025-07-17T09:17:08.5711442Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0107s] [ 86%] 2025-07-17T09:17:08.5711882Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0107s] [ 86%] 2025-07-17T09:17:08.5712307Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0113s] [ 86%] 2025-07-17T09:17:08.5712746Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0116s] [ 86%] 2025-07-17T09:17:08.5713183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0130s] [ 86%] 2025-07-17T09:17:08.5713627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0136s] [ 86%] 2025-07-17T09:17:08.5714116Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0137s] [ 86%] 2025-07-17T09:17:08.5714557Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0135s] [ 86%] 2025-07-17T09:17:08.5714986Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0141s] [ 86%] 2025-07-17T09:17:08.5715426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0145s] [ 86%] 2025-07-17T09:17:08.5715928Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0052s] [ 86%] 2025-07-17T09:17:08.5716437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0052s] [ 86%] 2025-07-17T09:17:08.5716871Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0052s] [ 86%] 2025-07-17T09:17:08.5717435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0052s] [ 86%] 2025-07-17T09:17:08.5717877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0052s] [ 86%] 2025-07-17T09:17:08.5718310Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0052s] [ 86%] 2025-07-17T09:17:08.5718751Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0071s] [ 86%] 2025-07-17T09:17:08.5719205Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0071s] [ 86%] 2025-07-17T09:17:08.5719636Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0071s] [ 86%] 2025-07-17T09:17:08.5720073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0070s] [ 86%] 2025-07-17T09:17:08.5720507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0070s] [ 86%] 2025-07-17T09:17:08.5720944Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0071s] [ 86%] 2025-07-17T09:17:08.5721372Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0056s] [ 86%] 2025-07-17T09:17:08.5721811Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0056s] [ 86%] 2025-07-17T09:17:08.5722241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0055s] [ 86%] 2025-07-17T09:17:08.5722670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0056s] [ 86%] 2025-07-17T09:17:08.5723154Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0055s] [ 86%] 2025-07-17T09:17:08.5723631Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0054s] [ 86%] 2025-07-17T09:17:08.5724072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0073s] [ 86%] 2025-07-17T09:17:08.5724617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0073s] [ 86%] 2025-07-17T09:17:08.5725049Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0073s] [ 86%] 2025-07-17T09:17:08.5725485Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0073s] [ 86%] 2025-07-17T09:17:08.5725917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0072s] [ 86%] 2025-07-17T09:17:08.5726361Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0072s] [ 86%] 2025-07-17T09:17:08.5726793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0051s] [ 86%] 2025-07-17T09:17:08.5727234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0051s] [ 86%] 2025-07-17T09:17:08.5727660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0051s] [ 86%] 2025-07-17T09:17:08.5728096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0050s] [ 86%] 2025-07-17T09:17:08.5728520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0052s] [ 86%] 2025-07-17T09:17:08.5728955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0050s] [ 86%] 2025-07-17T09:17:08.5729389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0069s] [ 86%] 2025-07-17T09:17:08.5729832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0067s] [ 86%] 2025-07-17T09:17:08.5730322Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0067s] [ 86%] 2025-07-17T09:17:08.5730805Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0067s] [ 86%] 2025-07-17T09:17:08.5731236Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0066s] [ 86%] 2025-07-17T09:17:08.5731784Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0066s] [ 86%] 2025-07-17T09:17:08.5732224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0054s] [ 86%] 2025-07-17T09:17:08.5732661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0053s] [ 86%] 2025-07-17T09:17:08.5733085Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0052s] [ 86%] 2025-07-17T09:17:08.5733519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0052s] [ 86%] 2025-07-17T09:17:08.5733943Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0051s] [ 86%] 2025-07-17T09:17:08.5734381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0052s] [ 86%] 2025-07-17T09:17:08.5734808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0071s] [ 86%] 2025-07-17T09:17:08.5735249Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0070s] [ 86%] 2025-07-17T09:17:08.5735676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0070s] [ 86%] 2025-07-17T09:17:08.5736112Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0070s] [ 86%] 2025-07-17T09:17:08.5736543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0068s] [ 86%] 2025-07-17T09:17:08.5736973Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0068s] [ 86%] 2025-07-17T09:17:08.5737463Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0787s] [ 86%] 2025-07-17T09:17:08.5737964Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0097s] [ 87%] 2025-07-17T09:17:08.5738397Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0095s] [ 87%] 2025-07-17T09:17:08.5738949Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0095s] [ 87%] 2025-07-17T09:17:08.5739389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0103s] [ 87%] 2025-07-17T09:17:08.5739830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0105s] [ 87%] 2025-07-17T09:17:08.5740264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0118s] [ 87%] 2025-07-17T09:17:08.5740712Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0120s] [ 87%] 2025-07-17T09:17:08.5741143Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0122s] [ 87%] 2025-07-17T09:17:08.5741582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0122s] [ 87%] 2025-07-17T09:17:08.5742011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0127s] [ 87%] 2025-07-17T09:17:08.5742449Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0132s] [ 87%] 2025-07-17T09:17:08.5742878Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0113s] [ 87%] 2025-07-17T09:17:08.5743320Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0098s] [ 87%] 2025-07-17T09:17:08.5743750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0107s] [ 87%] 2025-07-17T09:17:08.5744180Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0099s] [ 87%] 2025-07-17T09:17:08.5744665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0102s] [ 87%] 2025-07-17T09:17:08.5745142Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0082s] [ 87%] 2025-07-17T09:17:08.5745633Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0121s] [ 87%] 2025-07-17T09:17:08.5746121Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0126s] [ 87%] 2025-07-17T09:17:08.5746555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0127s] [ 87%] 2025-07-17T09:17:08.5746993Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0126s] [ 87%] 2025-07-17T09:17:08.5747423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0130s] [ 87%] 2025-07-17T09:17:08.5747862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0135s] [ 87%] 2025-07-17T09:17:08.5748293Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0050s] [ 87%] 2025-07-17T09:17:08.5748739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 87%] 2025-07-17T09:17:08.5749170Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0045s] [ 87%] 2025-07-17T09:17:08.5749605Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 87%] 2025-07-17T09:17:08.5750033Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0043s] [ 87%] 2025-07-17T09:17:08.5750468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0043s] [ 87%] 2025-07-17T09:17:08.5750919Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0051s] [ 87%] 2025-07-17T09:17:08.5751361Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0050s] [ 87%] 2025-07-17T09:17:08.5751856Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0050s] [ 87%] 2025-07-17T09:17:08.5752340Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0050s] [ 87%] 2025-07-17T09:17:08.5752875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0049s] [ 87%] 2025-07-17T09:17:08.5753312Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0049s] [ 87%] 2025-07-17T09:17:08.5753751Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 87%] 2025-07-17T09:17:08.5754253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 87%] 2025-07-17T09:17:08.5754697Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0046s] [ 87%] 2025-07-17T09:17:08.5755131Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0046s] [ 87%] 2025-07-17T09:17:08.5755561Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0044s] [ 87%] 2025-07-17T09:17:08.5755994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0044s] [ 87%] 2025-07-17T09:17:08.5756427Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0051s] [ 87%] 2025-07-17T09:17:08.5756874Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0051s] [ 87%] 2025-07-17T09:17:08.5757298Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0051s] [ 87%] 2025-07-17T09:17:08.5757735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0051s] [ 87%] 2025-07-17T09:17:08.5758165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0050s] [ 87%] 2025-07-17T09:17:08.5758597Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0049s] [ 87%] 2025-07-17T09:17:08.5759097Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 87%] 2025-07-17T09:17:08.5759589Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 87%] 2025-07-17T09:17:08.5760145Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 87%] 2025-07-17T09:17:08.5760579Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 87%] 2025-07-17T09:17:08.5761009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5761445Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 87%] 2025-07-17T09:17:08.5761882Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 87%] 2025-07-17T09:17:08.5762329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 87%] 2025-07-17T09:17:08.5762756Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 87%] 2025-07-17T09:17:08.5763190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 87%] 2025-07-17T09:17:08.5763621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 87%] 2025-07-17T09:17:08.5764061Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5764492Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5764928Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 87%] 2025-07-17T09:17:08.5765364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 87%] 2025-07-17T09:17:08.5765858Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 87%] 2025-07-17T09:17:08.5766339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 87%] 2025-07-17T09:17:08.5766763Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 87%] 2025-07-17T09:17:08.5767300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 87%] 2025-07-17T09:17:08.5767734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 87%] 2025-07-17T09:17:08.5768168Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0038s] [ 87%] 2025-07-17T09:17:08.5768601Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 87%] 2025-07-17T09:17:08.5769029Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5769463Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5769897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 87%] 2025-07-17T09:17:08.5770336Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 87%] 2025-07-17T09:17:08.5770759Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 87%] 2025-07-17T09:17:08.5771195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 87%] 2025-07-17T09:17:08.5771617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 87%] 2025-07-17T09:17:08.5772048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 87%] 2025-07-17T09:17:08.5772488Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 87%] 2025-07-17T09:17:08.5773103Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 87%] 2025-07-17T09:17:08.5773582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 87%] 2025-07-17T09:17:08.5774009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5774562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5774999Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 87%] 2025-07-17T09:17:08.5775447Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 87%] 2025-07-17T09:17:08.5775889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 87%] 2025-07-17T09:17:08.5776317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 87%] 2025-07-17T09:17:08.5776751Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 87%] 2025-07-17T09:17:08.5777177Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 87%] 2025-07-17T09:17:08.5777612Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 87%] 2025-07-17T09:17:08.5778045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 87%] 2025-07-17T09:17:08.5778485Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 87%] 2025-07-17T09:17:08.5778908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 87%] 2025-07-17T09:17:08.5779339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 87%] 2025-07-17T09:17:08.5779776Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5780258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5780742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 87%] 2025-07-17T09:17:08.5781178Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 87%] 2025-07-17T09:17:08.5781719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0034s] [ 87%] 2025-07-17T09:17:08.5782153Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5782584Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 87%] 2025-07-17T09:17:08.5783015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 87%] 2025-07-17T09:17:08.5783449Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 87%] 2025-07-17T09:17:08.5783892Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 87%] 2025-07-17T09:17:08.5784318Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 87%] 2025-07-17T09:17:08.5784755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 87%] 2025-07-17T09:17:08.5785184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0039s] [ 87%] 2025-07-17T09:17:08.5785623Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 87%] 2025-07-17T09:17:08.5786052Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 87%] 2025-07-17T09:17:08.5786489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5786921Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5787403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5787883Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 87%] 2025-07-17T09:17:08.5788308Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 87%] 2025-07-17T09:17:08.5788868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 87%] 2025-07-17T09:17:08.5789309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 87%] 2025-07-17T09:17:08.5789736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0045s] [ 87%] 2025-07-17T09:17:08.5790166Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 88%] 2025-07-17T09:17:08.5790591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0045s] [ 88%] 2025-07-17T09:17:08.5791025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 88%] 2025-07-17T09:17:08.5791462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.2293s] [ 88%] 2025-07-17T09:17:08.5791909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1268s] [ 88%] 2025-07-17T09:17:08.5792347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.1116s] [ 88%] 2025-07-17T09:17:08.5792796Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.1358s] [ 88%] 2025-07-17T09:17:08.5793224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.1329s] [ 88%] 2025-07-17T09:17:08.5793664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.1508s] [ 88%] 2025-07-17T09:17:08.5794221Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.1431s] [ 88%] 2025-07-17T09:17:08.5794735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1672s] [ 88%] 2025-07-17T09:17:08.5795227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.1687s] [ 88%] 2025-07-17T09:17:08.5795661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.1673s] [ 88%] 2025-07-17T09:17:08.5796226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.1530s] [ 88%] 2025-07-17T09:17:08.5796665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.1617s] [ 88%] 2025-07-17T09:17:08.5797108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.1440s] [ 88%] 2025-07-17T09:17:08.5797550Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1461s] [ 88%] 2025-07-17T09:17:08.5797981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.1365s] [ 88%] 2025-07-17T09:17:08.5798413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.1425s] [ 88%] 2025-07-17T09:17:08.5798842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0786s] [ 88%] 2025-07-17T09:17:08.5799284Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0158s] [ 88%] 2025-07-17T09:17:08.5799718Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0216s] [ 88%] 2025-07-17T09:17:08.5800163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1323s] [ 88%] 2025-07-17T09:17:08.5800592Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.1427s] [ 88%] 2025-07-17T09:17:08.5801035Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.1274s] [ 88%] 2025-07-17T09:17:08.5801471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.1087s] [ 88%] 2025-07-17T09:17:08.5801958Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.1309s] [ 88%] 2025-07-17T09:17:08.5802448Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.1081s] [ 88%] 2025-07-17T09:17:08.5802936Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0091s] [ 88%] 2025-07-17T09:17:08.5803426Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0710s] [ 88%] 2025-07-17T09:17:08.5803861Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0717s] [ 88%] 2025-07-17T09:17:08.5804294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0856s] [ 88%] 2025-07-17T09:17:08.5804737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0893s] [ 88%] 2025-07-17T09:17:08.5805172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0919s] [ 88%] 2025-07-17T09:17:08.5805614Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1155s] [ 88%] 2025-07-17T09:17:08.5806043Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0861s] [ 88%] 2025-07-17T09:17:08.5806483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.1022s] [ 88%] 2025-07-17T09:17:08.5806909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.1138s] [ 88%] 2025-07-17T09:17:08.5807350Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0925s] [ 88%] 2025-07-17T09:17:08.5807777Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0445s] [ 88%] 2025-07-17T09:17:08.5808223Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0311s] [ 88%] 2025-07-17T09:17:08.5808659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0312s] [ 88%] 2025-07-17T09:17:08.5809138Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0320s] [ 88%] 2025-07-17T09:17:08.5809620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0363s] [ 88%] 2025-07-17T09:17:08.5810153Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0419s] [ 88%] 2025-07-17T09:17:08.5810595Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.1243s] [ 88%] 2025-07-17T09:17:08.5811032Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1120s] [ 88%] 2025-07-17T09:17:08.5811466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.1439s] [ 88%] 2025-07-17T09:17:08.5811905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.1650s] [ 88%] 2025-07-17T09:17:08.5812332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.1473s] [ 88%] 2025-07-17T09:17:08.5812771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.1473s] [ 88%] 2025-07-17T09:17:08.5813207Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.1062s] [ 88%] 2025-07-17T09:17:08.5813659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0909s] [ 88%] 2025-07-17T09:17:08.5814089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.1060s] [ 88%] 2025-07-17T09:17:08.5814526Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.1266s] [ 88%] 2025-07-17T09:17:08.5814953Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0942s] [ 88%] 2025-07-17T09:17:08.5815390Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.1018s] [ 88%] 2025-07-17T09:17:08.5815832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.1546s] [ 88%] 2025-07-17T09:17:08.5816323Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1454s] [ 88%] 2025-07-17T09:17:08.5816804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0118s] [ 88%] 2025-07-17T09:17:08.5817366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.1332s] [ 88%] 2025-07-17T09:17:08.5817803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.1440s] [ 88%] 2025-07-17T09:17:08.5818235Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.1440s] [ 88%] 2025-07-17T09:17:08.5818672Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.1070s] [ 88%] 2025-07-17T09:17:08.5819109Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1009s] [ 88%] 2025-07-17T09:17:08.5819537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.1070s] [ 88%] 2025-07-17T09:17:08.5819972Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.1071s] [ 88%] 2025-07-17T09:17:08.5820396Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.1022s] [ 88%] 2025-07-17T09:17:08.5820836Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.1261s] [ 88%] 2025-07-17T09:17:08.5821267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.1318s] [ 88%] 2025-07-17T09:17:08.5821709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1522s] [ 88%] 2025-07-17T09:17:08.5822132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.1455s] [ 88%] 2025-07-17T09:17:08.5822571Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.1352s] [ 88%] 2025-07-17T09:17:08.5823053Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.1455s] [ 88%] 2025-07-17T09:17:08.5823552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.1454s] [ 88%] 2025-07-17T09:17:08.5823998Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.1477s] [ 88%] 2025-07-17T09:17:08.5824546Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1493s] [ 88%] 2025-07-17T09:17:08.5824985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.1510s] [ 88%] 2025-07-17T09:17:08.5825418Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.1386s] [ 88%] 2025-07-17T09:17:08.5825853Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.1591s] [ 88%] 2025-07-17T09:17:08.5826291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.1470s] [ 88%] 2025-07-17T09:17:08.5826729Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.1616s] [ 88%] 2025-07-17T09:17:08.5827171Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1667s] [ 88%] 2025-07-17T09:17:08.5827600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.1769s] [ 88%] 2025-07-17T09:17:08.5828038Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.1670s] [ 88%] 2025-07-17T09:17:08.5828466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.1684s] [ 88%] 2025-07-17T09:17:08.5828904Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.1841s] [ 88%] 2025-07-17T09:17:08.5829339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.1859s] [ 88%] 2025-07-17T09:17:08.5829773Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1267s] [ 88%] 2025-07-17T09:17:08.5830253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.1377s] [ 88%] 2025-07-17T09:17:08.5830738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.1149s] [ 88%] 2025-07-17T09:17:08.5831172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0670s] [ 88%] 2025-07-17T09:17:08.5831703Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0148s] [ 88%] 2025-07-17T09:17:08.5832148Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.1602s] [ 88%] 2025-07-17T09:17:08.5832587Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.1684s] [ 88%] 2025-07-17T09:17:08.5833016Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.1693s] [ 88%] 2025-07-17T09:17:08.5833454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.1464s] [ 88%] 2025-07-17T09:17:08.5833966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.2412s] [ 88%] 2025-07-17T09:17:08.5834409Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.1691s] [ 88%] 2025-07-17T09:17:08.5834841Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 88%] 2025-07-17T09:17:08.5835291Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 88%] 2025-07-17T09:17:08.5835718Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0051s] [ 88%] 2025-07-17T09:17:08.5836154Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0043s] [ 88%] 2025-07-17T09:17:08.5836581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0042s] [ 88%] 2025-07-17T09:17:08.5837011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 88%] 2025-07-17T09:17:08.5837514Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 88%] 2025-07-17T09:17:08.5838009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 88%] 2025-07-17T09:17:08.5838436Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0047s] [ 88%] 2025-07-17T09:17:08.5838998Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0047s] [ 88%] 2025-07-17T09:17:08.5839435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0046s] [ 88%] 2025-07-17T09:17:08.5839865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0046s] [ 88%] 2025-07-17T09:17:08.5840294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 88%] 2025-07-17T09:17:08.5840738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 88%] 2025-07-17T09:17:08.5841162Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0102s] [ 88%] 2025-07-17T09:17:08.5841608Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0044s] [ 88%] 2025-07-17T09:17:08.5842032Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0043s] [ 88%] 2025-07-17T09:17:08.5842466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0068s] [ 89%] 2025-07-17T09:17:08.5842905Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0281s] [ 89%] 2025-07-17T09:17:08.5843350Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0052s] [ 89%] 2025-07-17T09:17:08.5843771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0133s] [ 89%] 2025-07-17T09:17:08.5844206Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0049s] [ 89%] 2025-07-17T09:17:08.5844697Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0145s] [ 89%] 2025-07-17T09:17:08.5845193Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0049s] [ 89%] 2025-07-17T09:17:08.5845624Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5846183Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5846617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5847043Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5847470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5847909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5848339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 89%] 2025-07-17T09:17:08.5848774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 89%] 2025-07-17T09:17:08.5849196Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0038s] [ 89%] 2025-07-17T09:17:08.5849634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 89%] 2025-07-17T09:17:08.5850054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 89%] 2025-07-17T09:17:08.5850489Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 89%] 2025-07-17T09:17:08.5850914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5851351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 89%] 2025-07-17T09:17:08.5851836Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5852313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5852738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 89%] 2025-07-17T09:17:08.5853268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 89%] 2025-07-17T09:17:08.5853704Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 89%] 2025-07-17T09:17:08.5854134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 89%] 2025-07-17T09:17:08.5854561Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 89%] 2025-07-17T09:17:08.5854998Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 89%] 2025-07-17T09:17:08.5855421Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 89%] 2025-07-17T09:17:08.5855851Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 89%] 2025-07-17T09:17:08.5856276Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5856716Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 89%] 2025-07-17T09:17:08.5857139Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 89%] 2025-07-17T09:17:08.5857573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 89%] 2025-07-17T09:17:08.5857998Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 89%] 2025-07-17T09:17:08.5858436Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 89%] 2025-07-17T09:17:08.5858922Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 89%] 2025-07-17T09:17:08.5859405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 89%] 2025-07-17T09:17:08.5859832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5860375Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5860808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5861239Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5861669Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 89%] 2025-07-17T09:17:08.5862117Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 89%] 2025-07-17T09:17:08.5862540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5862970Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 89%] 2025-07-17T09:17:08.5863386Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 89%] 2025-07-17T09:17:08.5863814Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 89%] 2025-07-17T09:17:08.5864241Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 89%] 2025-07-17T09:17:08.5864676Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 89%] 2025-07-17T09:17:08.5865092Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 89%] 2025-07-17T09:17:08.5865523Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 89%] 2025-07-17T09:17:08.5866011Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 89%] 2025-07-17T09:17:08.5866486Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 89%] 2025-07-17T09:17:08.5866920Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 89%] 2025-07-17T09:17:08.5867459Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5867895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0034s] [ 89%] 2025-07-17T09:17:08.5868321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5868747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5869172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 89%] 2025-07-17T09:17:08.5869614Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 89%] 2025-07-17T09:17:08.5870057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 89%] 2025-07-17T09:17:08.5870478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0038s] [ 89%] 2025-07-17T09:17:08.5870911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 89%] 2025-07-17T09:17:08.5871335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 89%] 2025-07-17T09:17:08.5871769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 89%] 2025-07-17T09:17:08.5872191Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5872624Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5873123Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5873605Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 89%] 2025-07-17T09:17:08.5874107Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0034s] [ 89%] 2025-07-17T09:17:08.5874658Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 89%] 2025-07-17T09:17:08.5875095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 89%] 2025-07-17T09:17:08.5875527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 89%] 2025-07-17T09:17:08.5875954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0045s] [ 89%] 2025-07-17T09:17:08.5876391Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 89%] 2025-07-17T09:17:08.5876818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0045s] [ 89%] 2025-07-17T09:17:08.5877251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_1024_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 89%] 2025-07-17T09:17:08.5877685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 89%] 2025-07-17T09:17:08.5878134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 89%] 2025-07-17T09:17:08.5878562Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 89%] 2025-07-17T09:17:08.5878999Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 89%] 2025-07-17T09:17:08.5879423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 89%] 2025-07-17T09:17:08.5879863Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 89%] 2025-07-17T09:17:08.5880362Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 89%] 2025-07-17T09:17:08.5880863Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 89%] 2025-07-17T09:17:08.5881301Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0046s] [ 89%] 2025-07-17T09:17:08.5881962Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 89%] 2025-07-17T09:17:08.5882405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0044s] [ 89%] 2025-07-17T09:17:08.5882842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0044s] [ 89%] 2025-07-17T09:17:08.5883280Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 89%] 2025-07-17T09:17:08.5883715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 89%] 2025-07-17T09:17:08.5884151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0040s] [ 89%] 2025-07-17T09:17:08.5884583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 89%] 2025-07-17T09:17:08.5885015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 89%] 2025-07-17T09:17:08.5885454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 89%] 2025-07-17T09:17:08.5885888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0046s] [ 89%] 2025-07-17T09:17:08.5886328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0047s] [ 89%] 2025-07-17T09:17:08.5886750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0047s] [ 89%] 2025-07-17T09:17:08.5887189Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0046s] [ 89%] 2025-07-17T09:17:08.5887679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0044s] [ 89%] 2025-07-17T09:17:08.5888163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0045s] [ 89%] 2025-07-17T09:17:08.5888605Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5889148Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5889584Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5890013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5890444Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5890874Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 89%] 2025-07-17T09:17:08.5891313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 89%] 2025-07-17T09:17:08.5891758Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 89%] 2025-07-17T09:17:08.5892182Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 89%] 2025-07-17T09:17:08.5892630Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 89%] 2025-07-17T09:17:08.5893062Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5893508Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 89%] 2025-07-17T09:17:08.5893935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 89%] 2025-07-17T09:17:08.5894375Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 90%] 2025-07-17T09:17:08.5894858Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 90%] 2025-07-17T09:17:08.5895335Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 90%] 2025-07-17T09:17:08.5895765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5896315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5896766Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 90%] 2025-07-17T09:17:08.5897203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 90%] 2025-07-17T09:17:08.5897633Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 90%] 2025-07-17T09:17:08.5898071Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 90%] 2025-07-17T09:17:08.5898507Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5898942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5899370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5899810Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 90%] 2025-07-17T09:17:08.5900238Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5900735Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 90%] 2025-07-17T09:17:08.5901180Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 90%] 2025-07-17T09:17:08.5901617Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 90%] 2025-07-17T09:17:08.5902118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5902604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5903096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5903573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5904006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5904439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5904872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 90%] 2025-07-17T09:17:08.5905306Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5905734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5906166Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5906588Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 90%] 2025-07-17T09:17:08.5907021Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 90%] 2025-07-17T09:17:08.5907448Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5907887Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5908313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5908752Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5909228Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5909706Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5910195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 90%] 2025-07-17T09:17:08.5910683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5911111Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5911542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 90%] 2025-07-17T09:17:08.5911978Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5912404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5912852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0040s] [ 90%] 2025-07-17T09:17:08.5913295Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 90%] 2025-07-17T09:17:08.5913721Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0040s] [ 90%] 2025-07-17T09:17:08.5914229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 90%] 2025-07-17T09:17:08.5914658Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0039s] [ 90%] 2025-07-17T09:17:08.5915095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 90%] 2025-07-17T09:17:08.5915521Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5915970Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5916471Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5916967Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5917454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5917941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5918375Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 90%] 2025-07-17T09:17:08.5918807Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 90%] 2025-07-17T09:17:08.5919235Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0041s] [ 90%] 2025-07-17T09:17:08.5919679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 90%] 2025-07-17T09:17:08.5920114Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0041s] [ 90%] 2025-07-17T09:17:08.5920549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 90%] 2025-07-17T09:17:08.5920985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5921424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5921847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 90%] 2025-07-17T09:17:08.5922280Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5922705Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5923143Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 90%] 2025-07-17T09:17:08.5923636Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5929450Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5930093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 90%] 2025-07-17T09:17:08.5930549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 90%] 2025-07-17T09:17:08.5930982Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5931424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5931854Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 90%] 2025-07-17T09:17:08.5932292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 90%] 2025-07-17T09:17:08.5932737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0043s] [ 90%] 2025-07-17T09:17:08.5933169Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5933599Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5934027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5934460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 90%] 2025-07-17T09:17:08.5934894Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 90%] 2025-07-17T09:17:08.5935338Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 90%] 2025-07-17T09:17:08.5935766Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 90%] 2025-07-17T09:17:08.5936259Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5936747Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5937284Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5937725Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 90%] 2025-07-17T09:17:08.5938149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 90%] 2025-07-17T09:17:08.5938584Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 90%] 2025-07-17T09:17:08.5939012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 90%] 2025-07-17T09:17:08.5939440Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 90%] 2025-07-17T09:17:08.5939865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5940309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5940737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5941172Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5941600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 90%] 2025-07-17T09:17:08.5942028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 90%] 2025-07-17T09:17:08.5942478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5942959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5943439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5943864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 90%] 2025-07-17T09:17:08.5944402Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 90%] 2025-07-17T09:17:08.5944835Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 90%] 2025-07-17T09:17:08.5945259Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5945696Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5946118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 90%] 2025-07-17T09:17:08.5946551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5946974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5947404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5947833Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5948273Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 90%] 2025-07-17T09:17:08.5948699Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 90%] 2025-07-17T09:17:08.5949124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 90%] 2025-07-17T09:17:08.5949549Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 90%] 2025-07-17T09:17:08.5950027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 90%] 2025-07-17T09:17:08.5950509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 90%] 2025-07-17T09:17:08.5950945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5951479Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 90%] 2025-07-17T09:17:08.5951923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 91%] 2025-07-17T09:17:08.5952347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 91%] 2025-07-17T09:17:08.5952777Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 91%] 2025-07-17T09:17:08.5953214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 91%] 2025-07-17T09:17:08.5953650Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 91%] 2025-07-17T09:17:08.5954152Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 91%] 2025-07-17T09:17:08.5954580Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 91%] 2025-07-17T09:17:08.5955002Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 91%] 2025-07-17T09:17:08.5955430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 91%] 2025-07-17T09:17:08.5955868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 91%] 2025-07-17T09:17:08.5956300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 91%] 2025-07-17T09:17:08.5956729Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 91%] 2025-07-17T09:17:08.5957227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 91%] 2025-07-17T09:17:08.5957730Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 91%] 2025-07-17T09:17:08.5958153Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 91%] 2025-07-17T09:17:08.5958720Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 91%] 2025-07-17T09:17:08.5959155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 91%] 2025-07-17T09:17:08.5959586Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 91%] 2025-07-17T09:17:08.5960015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 91%] 2025-07-17T09:17:08.5960439Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 91%] 2025-07-17T09:17:08.5960870Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 91%] 2025-07-17T09:17:08.5961302Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 91%] 2025-07-17T09:17:08.5961740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 91%] 2025-07-17T09:17:08.5962166Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 91%] 2025-07-17T09:17:08.5962600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 91%] 2025-07-17T09:17:08.5963027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 91%] 2025-07-17T09:17:08.5963456Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 91%] 2025-07-17T09:17:08.5963888Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 91%] 2025-07-17T09:17:08.5964369Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 91%] 2025-07-17T09:17:08.5964842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 91%] 2025-07-17T09:17:08.5965264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 91%] 2025-07-17T09:17:08.5965795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0032s] [ 91%] 2025-07-17T09:17:08.5966222Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 91%] 2025-07-17T09:17:08.5966654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 91%] 2025-07-17T09:17:08.5967093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 91%] 2025-07-17T09:17:08.5967517Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0038s] [ 91%] 2025-07-17T09:17:08.5967946Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 91%] 2025-07-17T09:17:08.5968366Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 91%] 2025-07-17T09:17:08.5968795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 91%] 2025-07-17T09:17:08.5969231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.7453s] [ 91%] 2025-07-17T09:17:08.5969677Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0926s] [ 91%] 2025-07-17T09:17:08.5970106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0797s] [ 91%] 2025-07-17T09:17:08.5970540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0948s] [ 91%] 2025-07-17T09:17:08.5970973Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0785s] [ 91%] 2025-07-17T09:17:08.5971453Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0653s] [ 91%] 2025-07-17T09:17:08.5971942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0846s] [ 91%] 2025-07-17T09:17:08.5972385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0853s] [ 91%] 2025-07-17T09:17:08.5972923Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0270s] [ 91%] 2025-07-17T09:17:08.5973359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0096s] [ 91%] 2025-07-17T09:17:08.5973791Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0098s] [ 91%] 2025-07-17T09:17:08.5974224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0708s] [ 91%] 2025-07-17T09:17:08.5974658Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0644s] [ 91%] 2025-07-17T09:17:08.5975102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0982s] [ 91%] 2025-07-17T09:17:08.5975530Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0722s] [ 91%] 2025-07-17T09:17:08.5975963Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0911s] [ 91%] 2025-07-17T09:17:08.5976389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0721s] [ 91%] 2025-07-17T09:17:08.5976824Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0906s] [ 91%] 2025-07-17T09:17:08.5977264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0464s] [ 91%] 2025-07-17T09:17:08.5977698Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0918s] [ 91%] 2025-07-17T09:17:08.5978130Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0664s] [ 91%] 2025-07-17T09:17:08.5978627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0914s] [ 91%] 2025-07-17T09:17:08.5979108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0773s] [ 91%] 2025-07-17T09:17:08.5979537Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0808s] [ 91%] 2025-07-17T09:17:08.5980084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 91%] 2025-07-17T09:17:08.5980524Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 91%] 2025-07-17T09:17:08.5980961Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0041s] [ 91%] 2025-07-17T09:17:08.5981394Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 91%] 2025-07-17T09:17:08.5981827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0039s] [ 91%] 2025-07-17T09:17:08.5982267Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 91%] 2025-07-17T09:17:08.5982700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0048s] [ 91%] 2025-07-17T09:17:08.5983142Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0048s] [ 91%] 2025-07-17T09:17:08.5983569Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0057s] [ 91%] 2025-07-17T09:17:08.5984004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0053s] [ 91%] 2025-07-17T09:17:08.5984435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0048s] [ 91%] 2025-07-17T09:17:08.5984865Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0049s] [ 91%] 2025-07-17T09:17:08.5985300Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 91%] 2025-07-17T09:17:08.5985803Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 91%] 2025-07-17T09:17:08.5986283Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0045s] [ 91%] 2025-07-17T09:17:08.5986708Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0043s] [ 91%] 2025-07-17T09:17:08.5987250Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0042s] [ 91%] 2025-07-17T09:17:08.5987685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 91%] 2025-07-17T09:17:08.5988124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0052s] [ 91%] 2025-07-17T09:17:08.5988564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0051s] [ 91%] 2025-07-17T09:17:08.5988996Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0053s] [ 91%] 2025-07-17T09:17:08.5989430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0057s] [ 91%] 2025-07-17T09:17:08.5989854Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0053s] [ 91%] 2025-07-17T09:17:08.5990289Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0051s] [ 91%] 2025-07-17T09:17:08.5990722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 91%] 2025-07-17T09:17:08.5991164Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 91%] 2025-07-17T09:17:08.5991600Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 91%] 2025-07-17T09:17:08.5992027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 91%] 2025-07-17T09:17:08.5992462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 91%] 2025-07-17T09:17:08.5992941Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 91%] 2025-07-17T09:17:08.5993430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0052s] [ 91%] 2025-07-17T09:17:08.5993869Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 91%] 2025-07-17T09:17:08.5994495Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0041s] [ 91%] 2025-07-17T09:17:08.5994925Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 91%] 2025-07-17T09:17:08.5995357Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0039s] [ 91%] 2025-07-17T09:17:08.5995788Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 91%] 2025-07-17T09:17:08.5996218Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 91%] 2025-07-17T09:17:08.5996655Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 91%] 2025-07-17T09:17:08.5997079Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 91%] 2025-07-17T09:17:08.5997509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0044s] [ 91%] 2025-07-17T09:17:08.5997934Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 91%] 2025-07-17T09:17:08.5998370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 91%] 2025-07-17T09:17:08.5998805Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 91%] 2025-07-17T09:17:08.5999236Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 91%] 2025-07-17T09:17:08.5999666Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0042s] [ 91%] 2025-07-17T09:17:08.6000151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 91%] 2025-07-17T09:17:08.6000637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0039s] [ 91%] 2025-07-17T09:17:08.6001066Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 91%] 2025-07-17T09:17:08.6001818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [1.2236s] [ 91%] 2025-07-17T09:17:08.6002260Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0606s] [ 91%] 2025-07-17T09:17:08.6002691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0222s] [ 91%] 2025-07-17T09:17:08.6003124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0108s] [ 91%] 2025-07-17T09:17:08.6003551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0507s] [ 91%] 2025-07-17T09:17:08.6003984Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0100s] [ 92%] 2025-07-17T09:17:08.6004414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0160s] [ 92%] 2025-07-17T09:17:08.6004852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0493s] [ 92%] 2025-07-17T09:17:08.6005279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0686s] [ 92%] 2025-07-17T09:17:08.6005717Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0604s] [ 92%] 2025-07-17T09:17:08.6006151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0452s] [ 92%] 2025-07-17T09:17:08.6006582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0847s] [ 92%] 2025-07-17T09:17:08.6007013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0618s] [ 92%] 2025-07-17T09:17:08.6007499Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0132s] [ 92%] 2025-07-17T09:17:08.6007975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0733s] [ 92%] 2025-07-17T09:17:08.6008454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0488s] [ 92%] 2025-07-17T09:17:08.6008946Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0462s] [ 92%] 2025-07-17T09:17:08.6009379Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0051s] [ 92%] 2025-07-17T09:17:08.6009817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0879s] [ 92%] 2025-07-17T09:17:08.6010259Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0562s] [ 92%] 2025-07-17T09:17:08.6010684Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0339s] [ 92%] 2025-07-17T09:17:08.6011119Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0403s] [ 92%] 2025-07-17T09:17:08.6011542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0339s] [ 92%] 2025-07-17T09:17:08.6011976Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0353s] [ 92%] 2025-07-17T09:17:08.6012401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 92%] 2025-07-17T09:17:08.6012839Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 92%] 2025-07-17T09:17:08.6013269Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 92%] 2025-07-17T09:17:08.6013695Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 92%] 2025-07-17T09:17:08.6014122Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 92%] 2025-07-17T09:17:08.6014603Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 92%] 2025-07-17T09:17:08.6015091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 92%] 2025-07-17T09:17:08.6015576Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 92%] 2025-07-17T09:17:08.6016055Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 92%] 2025-07-17T09:17:08.6016481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 92%] 2025-07-17T09:17:08.6016911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 92%] 2025-07-17T09:17:08.6017342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 92%] 2025-07-17T09:17:08.6017771Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6018201Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6018622Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6019051Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6019472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0032s] [ 92%] 2025-07-17T09:17:08.6019897Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 92%] 2025-07-17T09:17:08.6020331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 92%] 2025-07-17T09:17:08.6020764Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0039s] [ 92%] 2025-07-17T09:17:08.6021193Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 92%] 2025-07-17T09:17:08.6021672Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 92%] 2025-07-17T09:17:08.6022151Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 92%] 2025-07-17T09:17:08.6022629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 92%] 2025-07-17T09:17:08.6023108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 92%] 2025-07-17T09:17:08.6023540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 92%] 2025-07-17T09:17:08.6023967Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0039s] [ 92%] 2025-07-17T09:17:08.6024394Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 92%] 2025-07-17T09:17:08.6024813Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 92%] 2025-07-17T09:17:08.6025237Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 92%] 2025-07-17T09:17:08.6025661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 92%] 2025-07-17T09:17:08.6026098Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 92%] 2025-07-17T09:17:08.6026520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6026951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6027376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 92%] 2025-07-17T09:17:08.6027802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 92%] 2025-07-17T09:17:08.6028234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0031s] [ 92%] 2025-07-17T09:17:08.6028739Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 92%] 2025-07-17T09:17:08.6029214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 92%] 2025-07-17T09:17:08.6029633Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 92%] 2025-07-17T09:17:08.6030166Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 92%] 2025-07-17T09:17:08.6030593Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 92%] 2025-07-17T09:17:08.6031027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 92%] 2025-07-17T09:17:08.6031461Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 92%] 2025-07-17T09:17:08.6031878Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 92%] 2025-07-17T09:17:08.6032305Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 92%] 2025-07-17T09:17:08.6032722Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6033149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6033573Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 92%] 2025-07-17T09:17:08.6034067Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 92%] 2025-07-17T09:17:08.6034485Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0029s] [ 92%] 2025-07-17T09:17:08.6034912Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 92%] 2025-07-17T09:17:08.6035332Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0028s] [ 92%] 2025-07-17T09:17:08.6035820Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 92%] 2025-07-17T09:17:08.6036308Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6036736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6037309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6037737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6038161Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 92%] 2025-07-17T09:17:08.6038588Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 92%] 2025-07-17T09:17:08.6039012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 92%] 2025-07-17T09:17:08.6039442Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 92%] 2025-07-17T09:17:08.6039863Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 92%] 2025-07-17T09:17:08.6040286Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 92%] 2025-07-17T09:17:08.6040700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 92%] 2025-07-17T09:17:08.6041124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 92%] 2025-07-17T09:17:08.6041545Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 92%] 2025-07-17T09:17:08.6041976Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 92%] 2025-07-17T09:17:08.6042401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 92%] 2025-07-17T09:17:08.6042872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 92%] 2025-07-17T09:17:08.6043341Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6043766Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6044302Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 92%] 2025-07-17T09:17:08.6044736Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 92%] 2025-07-17T09:17:08.6045165Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 92%] 2025-07-17T09:17:08.6045594Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 92%] 2025-07-17T09:17:08.6046015Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0028s] [ 92%] 2025-07-17T09:17:08.6046443Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 92%] 2025-07-17T09:17:08.6046868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6047304Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6047728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 92%] 2025-07-17T09:17:08.6048157Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 92%] 2025-07-17T09:17:08.6048575Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 92%] 2025-07-17T09:17:08.6049002Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 92%] 2025-07-17T09:17:08.6049430Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0031s] [ 92%] 2025-07-17T09:17:08.6049912Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 92%] 2025-07-17T09:17:08.6050385Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 92%] 2025-07-17T09:17:08.6050804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 92%] 2025-07-17T09:17:08.6051337Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 92%] 2025-07-17T09:17:08.6051761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 92%] 2025-07-17T09:17:08.6052193Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 92%] 2025-07-17T09:17:08.6052619Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 92%] 2025-07-17T09:17:08.6053045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 92%] 2025-07-17T09:17:08.6053475Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 92%] 2025-07-17T09:17:08.6053891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 92%] 2025-07-17T09:17:08.6054316Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_103_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 92%] 2025-07-17T09:17:08.6054755Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0354s] [ 92%] 2025-07-17T09:17:08.6055207Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0147s] [ 93%] 2025-07-17T09:17:08.6055638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0148s] [ 93%] 2025-07-17T09:17:08.6056077Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0144s] [ 93%] 2025-07-17T09:17:08.6056509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0169s] [ 93%] 2025-07-17T09:17:08.6056997Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0171s] [ 93%] 2025-07-17T09:17:08.6057626Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0044s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6058237Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6058852Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6059405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0033s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6059960Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6060520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6060958Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0125s] [ 93%] 2025-07-17T09:17:08.6061403Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0149s] [ 93%] 2025-07-17T09:17:08.6061832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0150s] [ 93%] 2025-07-17T09:17:08.6062268Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0628s] [ 93%] 2025-07-17T09:17:08.6062707Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0725s] [ 93%] 2025-07-17T09:17:08.6063149Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0608s] [ 93%] 2025-07-17T09:17:08.6063710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0493s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6064323Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6064943Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6065542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6066150Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0143s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6066707Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6067142Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0085s] [ 93%] 2025-07-17T09:17:08.6067588Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0083s] [ 93%] 2025-07-17T09:17:08.6068017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0378s] [ 93%] 2025-07-17T09:17:08.6068451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0496s] [ 93%] 2025-07-17T09:17:08.6068878Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0503s] [ 93%] 2025-07-17T09:17:08.6069317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0420s] [ 93%] 2025-07-17T09:17:08.6069872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0330s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6070431Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6070987Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0292s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6071589Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6072199Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6072860Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6073294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0091s] [ 93%] 2025-07-17T09:17:08.6073734Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0329s] [ 93%] 2025-07-17T09:17:08.6074232Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0452s] [ 93%] 2025-07-17T09:17:08.6074664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0677s] [ 93%] 2025-07-17T09:17:08.6075092Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0545s] [ 93%] 2025-07-17T09:17:08.6075523Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0304s] [ 93%] 2025-07-17T09:17:08.6076073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0210s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6076629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6077174Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6077721Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6078272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0039s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6078886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6079389Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0082s] [ 93%] 2025-07-17T09:17:08.6079886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0081s] [ 93%] 2025-07-17T09:17:08.6080377Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0080s] [ 93%] 2025-07-17T09:17:08.6080808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0080s] [ 93%] 2025-07-17T09:17:08.6081239Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0081s] [ 93%] 2025-07-17T09:17:08.6081672Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0081s] [ 93%] 2025-07-17T09:17:08.6082229Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6082786Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6083339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6083890Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6084438Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6084985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6085423Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0089s] [ 93%] 2025-07-17T09:17:08.6085914Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0088s] [ 93%] 2025-07-17T09:17:08.6086394Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0095s] [ 93%] 2025-07-17T09:17:08.6086821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0088s] [ 93%] 2025-07-17T09:17:08.6087365Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0088s] [ 93%] 2025-07-17T09:17:08.6087794Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0088s] [ 93%] 2025-07-17T09:17:08.6088359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6088913Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6089460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6090010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6090552Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6091106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6091543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.2831s] [ 93%] 2025-07-17T09:17:08.6091984Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0139s] [ 93%] 2025-07-17T09:17:08.6092414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0349s] [ 93%] 2025-07-17T09:17:08.6092847Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0549s] [ 93%] 2025-07-17T09:17:08.6093359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0602s] [ 93%] 2025-07-17T09:17:08.6093841Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0508s] [ 93%] 2025-07-17T09:17:08.6094510Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0388s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6095073Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6095623Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0235s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6096178Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6096728Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6097289Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6097725Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0120s] [ 93%] 2025-07-17T09:17:08.6098163Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0500s] [ 93%] 2025-07-17T09:17:08.6098595Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0714s] [ 93%] 2025-07-17T09:17:08.6099022Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0956s] [ 93%] 2025-07-17T09:17:08.6099449Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0637s] [ 93%] 2025-07-17T09:17:08.6099877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0133s] [ 93%] 2025-07-17T09:17:08.6100483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0410s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6101102Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6101760Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6102309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6102855Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0191s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6103407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6103841Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0062s] [ 93%] 2025-07-17T09:17:08.6104288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0122s] [ 93%] 2025-07-17T09:17:08.6104718Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0085s] [ 93%] 2025-07-17T09:17:08.6105153Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0080s] [ 93%] 2025-07-17T09:17:08.6105583Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0364s] [ 93%] 2025-07-17T09:17:08.6106013Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0808s] [ 93%] 2025-07-17T09:17:08.6106567Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0496s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6107121Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6107724Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0982s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6108328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6108985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6109543Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6109975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0061s] [ 93%] 2025-07-17T09:17:08.6110413Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0525s] [ 93%] 2025-07-17T09:17:08.6110842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0815s] [ 93%] 2025-07-17T09:17:08.6111275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.1356s] [ 93%] 2025-07-17T09:17:08.6111704Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0726s] [ 93%] 2025-07-17T09:17:08.6112133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0730s] [ 93%] 2025-07-17T09:17:08.6112695Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0580s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6113245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6113795Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 93%] 2025-07-17T09:17:08.6114417Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6115028Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0542s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6115697Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6116273Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 94%] 2025-07-17T09:17:08.6116721Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 94%] 2025-07-17T09:17:08.6117148Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0039s] [ 94%] 2025-07-17T09:17:08.6117578Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 94%] 2025-07-17T09:17:08.6118009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 94%] 2025-07-17T09:17:08.6118435Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 94%] 2025-07-17T09:17:08.6118991Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6119540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6120088Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6120639Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6121184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6121737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6122224Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0042s] [ 94%] 2025-07-17T09:17:08.6122710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 94%] 2025-07-17T09:17:08.6123204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0042s] [ 94%] 2025-07-17T09:17:08.6123684Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 94%] 2025-07-17T09:17:08.6124115Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0039s] [ 94%] 2025-07-17T09:17:08.6124548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 94%] 2025-07-17T09:17:08.6125100Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6125652Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6126198Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6126740Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6127288Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6127840Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6128269Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 94%] 2025-07-17T09:17:08.6128713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 94%] 2025-07-17T09:17:08.6129190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0039s] [ 94%] 2025-07-17T09:17:08.6129673Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 94%] 2025-07-17T09:17:08.6130096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 94%] 2025-07-17T09:17:08.6130640Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 94%] 2025-07-17T09:17:08.6131202Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6131761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6132314Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6132860Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6133407Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6133959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6134388Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 94%] 2025-07-17T09:17:08.6134825Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 94%] 2025-07-17T09:17:08.6135252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0045s] [ 94%] 2025-07-17T09:17:08.6135684Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 94%] 2025-07-17T09:17:08.6136105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 94%] 2025-07-17T09:17:08.6136591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 94%] 2025-07-17T09:17:08.6137190Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6137844Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6138401Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6138946Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6139496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6140048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6140482Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.8542s] [ 94%] 2025-07-17T09:17:08.6140921Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0623s] [ 94%] 2025-07-17T09:17:08.6141347Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0799s] [ 94%] 2025-07-17T09:17:08.6141781Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0871s] [ 94%] 2025-07-17T09:17:08.6142208Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0859s] [ 94%] 2025-07-17T09:17:08.6142637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0836s] [ 94%] 2025-07-17T09:17:08.6143196Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0999s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6143805Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6144400Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0530s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6145072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6145620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0052s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6146168Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6146604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0107s] [ 94%] 2025-07-17T09:17:08.6147045Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0519s] [ 94%] 2025-07-17T09:17:08.6147468Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0793s] [ 94%] 2025-07-17T09:17:08.6147899Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.1665s] [ 94%] 2025-07-17T09:17:08.6148322Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0833s] [ 94%] 2025-07-17T09:17:08.6148749Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0055s] [ 94%] 2025-07-17T09:17:08.6149298Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0417s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6149853Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0082s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6150395Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6150994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0065s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6151590Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0583s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6152245Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6152685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0200s] [ 94%] 2025-07-17T09:17:08.6153133Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.2218s] [ 94%] 2025-07-17T09:17:08.6153561Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.2863s] [ 94%] 2025-07-17T09:17:08.6154066Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.2821s] [ 94%] 2025-07-17T09:17:08.6154497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.2803s] [ 94%] 2025-07-17T09:17:08.6154935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.2890s] [ 94%] 2025-07-17T09:17:08.6155488Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.1261s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6156050Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0021s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6156607Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0290s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6157155Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0082s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6157710Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6158321Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6158818Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0208s] [ 94%] 2025-07-17T09:17:08.6159378Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.2500s] [ 94%] 2025-07-17T09:17:08.6159812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.2309s] [ 94%] 2025-07-17T09:17:08.6160252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.2845s] [ 94%] 2025-07-17T09:17:08.6160675Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.1555s] [ 94%] 2025-07-17T09:17:08.6161113Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.2006s] [ 94%] 2025-07-17T09:17:08.6161668Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.1181s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6162227Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0023s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6162774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6163325Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6163877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0344s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6164424Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0019s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6164862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0138s] [ 94%] 2025-07-17T09:17:08.6165351Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0136s] [ 94%] 2025-07-17T09:17:08.6165830Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.1212s] [ 94%] 2025-07-17T09:17:08.6166315Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.1428s] [ 94%] 2025-07-17T09:17:08.6166789Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0639s] [ 94%] 2025-07-17T09:17:08.6167226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0943s] [ 94%] 2025-07-17T09:17:08.6167779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0860s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6168342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6168887Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0264s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6169441Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6170000Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6170551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 94%] 2025-07-17T09:17:08.6170987Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0153s] [ 94%] 2025-07-17T09:17:08.6171420Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0834s] [ 94%] 2025-07-17T09:17:08.6171853Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.1145s] [ 94%] 2025-07-17T09:17:08.6172329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0940s] [ 94%] 2025-07-17T09:17:08.6172819Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0977s] [ 94%] 2025-07-17T09:17:08.6173258Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.1432s] [ 95%] 2025-07-17T09:17:08.6173920Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.1844s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6174481Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6175025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6175577Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6176130Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0035s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6176678Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6177116Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0329s] [ 95%] 2025-07-17T09:17:08.6177555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.2356s] [ 95%] 2025-07-17T09:17:08.6177986Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.2038s] [ 95%] 2025-07-17T09:17:08.6178414Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.2109s] [ 95%] 2025-07-17T09:17:08.6178843Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.2369s] [ 95%] 2025-07-17T09:17:08.6179278Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.2038s] [ 95%] 2025-07-17T09:17:08.6179889Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.1378s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6180491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6181150Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0566s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6181715Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6182260Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6182811Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6183248Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0153s] [ 95%] 2025-07-17T09:17:08.6183680Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.2064s] [ 95%] 2025-07-17T09:17:08.6184106Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.2104s] [ 95%] 2025-07-17T09:17:08.6184534Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.2300s] [ 95%] 2025-07-17T09:17:08.6184960Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.2157s] [ 95%] 2025-07-17T09:17:08.6185391Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.2237s] [ 95%] 2025-07-17T09:17:08.6185942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.1665s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6186576Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0017s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6187169Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6187717Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6188364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0315s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6188916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6189358Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0418s] [ 95%] 2025-07-17T09:17:08.6189798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.2979s] [ 95%] 2025-07-17T09:17:08.6190233Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.2554s] [ 95%] 2025-07-17T09:17:08.6190670Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.3052s] [ 95%] 2025-07-17T09:17:08.6191099Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.2451s] [ 95%] 2025-07-17T09:17:08.6191527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.2766s] [ 95%] 2025-07-17T09:17:08.6192083Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.2273s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6192638Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0021s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6193191Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0795s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6193800Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6194460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0047s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6195076Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0018s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6195569Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0191s] [ 95%] 2025-07-17T09:17:08.6196004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.2780s] [ 95%] 2025-07-17T09:17:08.6196434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.2372s] [ 95%] 2025-07-17T09:17:08.6196864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.3866s] [ 95%] 2025-07-17T09:17:08.6197292Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.2153s] [ 95%] 2025-07-17T09:17:08.6197719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0210s] [ 95%] 2025-07-17T09:17:08.6198275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.1544s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6198833Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0026s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6199380Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0022s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6199931Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0022s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6200473Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0511s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6201105Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0020s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6201591Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0049s] [ 95%] 2025-07-17T09:17:08.6202027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0044s] [ 95%] 2025-07-17T09:17:08.6202560Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0043s] [ 95%] 2025-07-17T09:17:08.6202994Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0043s] [ 95%] 2025-07-17T09:17:08.6203420Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0044s] [ 95%] 2025-07-17T09:17:08.6203846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0044s] [ 95%] 2025-07-17T09:17:08.6204402Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6204953Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6205500Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6206048Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6206592Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6207141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6207567Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0047s] [ 95%] 2025-07-17T09:17:08.6208004Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0044s] [ 95%] 2025-07-17T09:17:08.6208485Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0046s] [ 95%] 2025-07-17T09:17:08.6208959Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0044s] [ 95%] 2025-07-17T09:17:08.6209438Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0045s] [ 95%] 2025-07-17T09:17:08.6209911Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0045s] [ 95%] 2025-07-17T09:17:08.6210462Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6211010Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6211555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6212104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6212644Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0016s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6213186Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6213613Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 95%] 2025-07-17T09:17:08.6214057Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 95%] 2025-07-17T09:17:08.6214486Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0034s] [ 95%] 2025-07-17T09:17:08.6214918Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 95%] 2025-07-17T09:17:08.6215399Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 95%] 2025-07-17T09:17:08.6215875Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 95%] 2025-07-17T09:17:08.6216432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6217096Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6217647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6218191Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6218742Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6219287Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6219716Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 95%] 2025-07-17T09:17:08.6220154Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 95%] 2025-07-17T09:17:08.6220574Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 95%] 2025-07-17T09:17:08.6220998Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 95%] 2025-07-17T09:17:08.6221421Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0032s] [ 95%] 2025-07-17T09:17:08.6221846Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 95%] 2025-07-17T09:17:08.6222392Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6223002Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6223596Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6224252Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6224797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6225346Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6225778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 95%] 2025-07-17T09:17:08.6226210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 95%] 2025-07-17T09:17:08.6226630Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 95%] 2025-07-17T09:17:08.6227056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 95%] 2025-07-17T09:17:08.6227478Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 95%] 2025-07-17T09:17:08.6227906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 95%] 2025-07-17T09:17:08.6228451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6229000Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6229542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6230211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0015s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6230802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6231456Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 95%] 2025-07-17T09:17:08.6231891Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 95%] 2025-07-17T09:17:08.6232328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 96%] 2025-07-17T09:17:08.6232751Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0039s] [ 96%] 2025-07-17T09:17:08.6233184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 96%] 2025-07-17T09:17:08.6233604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 96%] 2025-07-17T09:17:08.6234092Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 96%] 2025-07-17T09:17:08.6234635Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6235203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6235748Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6236294Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6236838Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6237437Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6237957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0043s] [ 96%] 2025-07-17T09:17:08.6238542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0041s] [ 96%] 2025-07-17T09:17:08.6238974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0040s] [ 96%] 2025-07-17T09:17:08.6239406Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0039s] [ 96%] 2025-07-17T09:17:08.6239827Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 96%] 2025-07-17T09:17:08.6240264Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 96%] 2025-07-17T09:17:08.6240815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6241374Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6241916Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6242460Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6243014Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6243554Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6243985Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 96%] 2025-07-17T09:17:08.6244467Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 96%] 2025-07-17T09:17:08.6244935Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0043s] [ 96%] 2025-07-17T09:17:08.6245362Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0041s] [ 96%] 2025-07-17T09:17:08.6245892Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 96%] 2025-07-17T09:17:08.6246327Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0044s] [ 96%] 2025-07-17T09:17:08.6246877Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6247432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6247981Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6248540Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 SKIPPED [0.0013s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6249092Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 SKIPPED [0.0014s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6249633Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_2048_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 SKIPPED [0.0012s] (Will call _fill_mem_eff_dropout_mask with too many threads!) [ 96%] 2025-07-17T09:17:08.6250075Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 96%] 2025-07-17T09:17:08.6250514Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6250950Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6251382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 96%] 2025-07-17T09:17:08.6251868Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 96%] 2025-07-17T09:17:08.6252359Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 96%] 2025-07-17T09:17:08.6252917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 96%] 2025-07-17T09:17:08.6253363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 96%] 2025-07-17T09:17:08.6253793Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0040s] [ 96%] 2025-07-17T09:17:08.6254231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 96%] 2025-07-17T09:17:08.6254659Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0040s] [ 96%] 2025-07-17T09:17:08.6255091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 96%] 2025-07-17T09:17:08.6255519Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6255952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6256381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6256804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6257230Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0035s] [ 96%] 2025-07-17T09:17:08.6257654Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 96%] 2025-07-17T09:17:08.6258094Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 96%] 2025-07-17T09:17:08.6258532Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 96%] 2025-07-17T09:17:08.6259006Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0043s] [ 96%] 2025-07-17T09:17:08.6259484Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 96%] 2025-07-17T09:17:08.6260008Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0041s] [ 96%] 2025-07-17T09:17:08.6260450Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 96%] 2025-07-17T09:17:08.6260876Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 96%] 2025-07-17T09:17:08.6261313Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 96%] 2025-07-17T09:17:08.6261737Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 96%] 2025-07-17T09:17:08.6262170Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 96%] 2025-07-17T09:17:08.6262592Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 96%] 2025-07-17T09:17:08.6263017Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 96%] 2025-07-17T09:17:08.6263451Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6263884Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 96%] 2025-07-17T09:17:08.6264309Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6264732Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6265158Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 96%] 2025-07-17T09:17:08.6265647Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 96%] 2025-07-17T09:17:08.6266121Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 96%] 2025-07-17T09:17:08.6266556Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 96%] 2025-07-17T09:17:08.6267095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 96%] 2025-07-17T09:17:08.6267527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 96%] 2025-07-17T09:17:08.6267946Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 96%] 2025-07-17T09:17:08.6268376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 96%] 2025-07-17T09:17:08.6268798Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 96%] 2025-07-17T09:17:08.6269234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6269660Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0038s] [ 96%] 2025-07-17T09:17:08.6270084Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 96%] 2025-07-17T09:17:08.6270518Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6270940Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6271376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 96%] 2025-07-17T09:17:08.6271807Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 96%] 2025-07-17T09:17:08.6272234Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 96%] 2025-07-17T09:17:08.6272663Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 96%] 2025-07-17T09:17:08.6273153Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 96%] 2025-07-17T09:17:08.6273634Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 96%] 2025-07-17T09:17:08.6274244Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 96%] 2025-07-17T09:17:08.6274685Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 96%] 2025-07-17T09:17:08.6275109Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 96%] 2025-07-17T09:17:08.6275542Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 96%] 2025-07-17T09:17:08.6275965Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 96%] 2025-07-17T09:17:08.6276392Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 96%] 2025-07-17T09:17:08.6276821Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 96%] 2025-07-17T09:17:08.6277251Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 96%] 2025-07-17T09:17:08.6277680Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 96%] 2025-07-17T09:17:08.6278104Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 96%] 2025-07-17T09:17:08.6278527Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0032s] [ 96%] 2025-07-17T09:17:08.6278949Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 96%] 2025-07-17T09:17:08.6279381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6279811Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 96%] 2025-07-17T09:17:08.6280353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 96%] 2025-07-17T09:17:08.6280862Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 96%] 2025-07-17T09:17:08.6281402Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 96%] 2025-07-17T09:17:08.6281832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 96%] 2025-07-17T09:17:08.6282261Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 96%] 2025-07-17T09:17:08.6282700Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 96%] 2025-07-17T09:17:08.6283124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 96%] 2025-07-17T09:17:08.6283555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 96%] 2025-07-17T09:17:08.6283979Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 96%] 2025-07-17T09:17:08.6284404Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 96%] 2025-07-17T09:17:08.6284839Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 96%] 2025-07-17T09:17:08.6285274Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 96%] 2025-07-17T09:17:08.6285702Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0039s] [ 96%] 2025-07-17T09:17:08.6286132Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 97%] 2025-07-17T09:17:08.6286564Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0038s] [ 97%] 2025-07-17T09:17:08.6286988Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 97%] 2025-07-17T09:17:08.6287466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6287952Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6288477Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6288909Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6289330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0033s] [ 97%] 2025-07-17T09:17:08.6289757Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 97%] 2025-07-17T09:17:08.6290186Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 97%] 2025-07-17T09:17:08.6290620Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0043s] [ 97%] 2025-07-17T09:17:08.6291044Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0042s] [ 97%] 2025-07-17T09:17:08.6291470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0042s] [ 97%] 2025-07-17T09:17:08.6291895Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0038s] [ 97%] 2025-07-17T09:17:08.6292319Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_1024_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 97%] 2025-07-17T09:17:08.6292752Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6293184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 97%] 2025-07-17T09:17:08.6293607Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6294030Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6294503Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6294986Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 97%] 2025-07-17T09:17:08.6295466Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6295955Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6296382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6296815Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6297233Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6297663Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6298093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 97%] 2025-07-17T09:17:08.6298520Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 97%] 2025-07-17T09:17:08.6298942Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 97%] 2025-07-17T09:17:08.6299381Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 97%] 2025-07-17T09:17:08.6299804Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6300231Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6300665Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 97%] 2025-07-17T09:17:08.6301098Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 97%] 2025-07-17T09:17:08.6301581Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0037s] [ 97%] 2025-07-17T09:17:08.6302056Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 97%] 2025-07-17T09:17:08.6302525Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 97%] 2025-07-17T09:17:08.6303007Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6303434Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6303867Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6304289Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6304713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6305134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 97%] 2025-07-17T09:17:08.6305558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 97%] 2025-07-17T09:17:08.6305989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6306420Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6306842Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6307265Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6307693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 97%] 2025-07-17T09:17:08.6308117Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0040s] [ 97%] 2025-07-17T09:17:08.6308615Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6309093Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6309558Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6310038Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6310456Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6310879Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6311302Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 97%] 2025-07-17T09:17:08.6311738Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6312154Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 97%] 2025-07-17T09:17:08.6312585Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6313009Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6313432Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6313859Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6314356Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6314780Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6315203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6315719Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6316214Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 97%] 2025-07-17T09:17:08.6316642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6317205Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6317625Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6318054Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6318472Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6318902Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 97%] 2025-07-17T09:17:08.6319328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 97%] 2025-07-17T09:17:08.6319758Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6320179Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6320603Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6321024Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6321446Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6321872Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 97%] 2025-07-17T09:17:08.6322298Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6322769Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 97%] 2025-07-17T09:17:08.6323238Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 97%] 2025-07-17T09:17:08.6323664Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6324197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6324625Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6325060Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6325474Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6325902Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6326320Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 97%] 2025-07-17T09:17:08.6326750Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 97%] 2025-07-17T09:17:08.6327184Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6327621Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6328064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6328493Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6328917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6329343Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 97%] 2025-07-17T09:17:08.6329817Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 97%] 2025-07-17T09:17:08.6330290Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 97%] 2025-07-17T09:17:08.6330713Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0032s] [ 97%] 2025-07-17T09:17:08.6331253Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 97%] 2025-07-17T09:17:08.6331681Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6332108Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0031s] [ 97%] 2025-07-17T09:17:08.6332530Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 97%] 2025-07-17T09:17:08.6332966Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 97%] 2025-07-17T09:17:08.6333382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 97%] 2025-07-17T09:17:08.6333810Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 97%] 2025-07-17T09:17:08.6334235Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6334661Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_103_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 97%] 2025-07-17T09:17:08.6335095Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 97%] 2025-07-17T09:17:08.6335528Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 97%] 2025-07-17T09:17:08.6335951Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 97%] 2025-07-17T09:17:08.6336380Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 97%] 2025-07-17T09:17:08.6336858Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0040s] [ 97%] 2025-07-17T09:17:08.6337339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0039s] [ 98%] 2025-07-17T09:17:08.6337778Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 98%] 2025-07-17T09:17:08.6338394Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0046s] [ 98%] 2025-07-17T09:17:08.6338823Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0045s] [ 98%] 2025-07-17T09:17:08.6339260Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 98%] 2025-07-17T09:17:08.6339689Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0046s] [ 98%] 2025-07-17T09:17:08.6340124Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0046s] [ 98%] 2025-07-17T09:17:08.6340548Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0039s] [ 98%] 2025-07-17T09:17:08.6340984Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 98%] 2025-07-17T09:17:08.6341408Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0038s] [ 98%] 2025-07-17T09:17:08.6341836Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 98%] 2025-07-17T09:17:08.6342275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0038s] [ 98%] 2025-07-17T09:17:08.6342709Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 98%] 2025-07-17T09:17:08.6343146Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0045s] [ 98%] 2025-07-17T09:17:08.6343582Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0045s] [ 98%] 2025-07-17T09:17:08.6344064Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0045s] [ 98%] 2025-07-17T09:17:08.6344534Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0045s] [ 98%] 2025-07-17T09:17:08.6344958Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0045s] [ 98%] 2025-07-17T09:17:08.6345494Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0045s] [ 98%] 2025-07-17T09:17:08.6345924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 98%] 2025-07-17T09:17:08.6346360Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 98%] 2025-07-17T09:17:08.6346781Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 98%] 2025-07-17T09:17:08.6347211Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 98%] 2025-07-17T09:17:08.6347629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 98%] 2025-07-17T09:17:08.6348062Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 98%] 2025-07-17T09:17:08.6348493Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6348927Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 98%] 2025-07-17T09:17:08.6349353Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6349785Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6350209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 98%] 2025-07-17T09:17:08.6350636Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 98%] 2025-07-17T09:17:08.6351115Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 98%] 2025-07-17T09:17:08.6351592Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0032s] [ 98%] 2025-07-17T09:17:08.6352020Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 98%] 2025-07-17T09:17:08.6352555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 98%] 2025-07-17T09:17:08.6352976Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0032s] [ 98%] 2025-07-17T09:17:08.6353406Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 98%] 2025-07-17T09:17:08.6353834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0038s] [ 98%] 2025-07-17T09:17:08.6354348Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0038s] [ 98%] 2025-07-17T09:17:08.6354777Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0038s] [ 98%] 2025-07-17T09:17:08.6355209Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 98%] 2025-07-17T09:17:08.6355627Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6356058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6356497Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 98%] 2025-07-17T09:17:08.6356934Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 98%] 2025-07-17T09:17:08.6357364Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 98%] 2025-07-17T09:17:08.6357790Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 98%] 2025-07-17T09:17:08.6358279Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 98%] 2025-07-17T09:17:08.6358765Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 98%] 2025-07-17T09:17:08.6359197Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 98%] 2025-07-17T09:17:08.6359780Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 98%] 2025-07-17T09:17:08.6360204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6360635Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6361055Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6361490Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6361912Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0033s] [ 98%] 2025-07-17T09:17:08.6362342Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 98%] 2025-07-17T09:17:08.6362762Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0033s] [ 98%] 2025-07-17T09:17:08.6363188Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0033s] [ 98%] 2025-07-17T09:17:08.6363609Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0032s] [ 98%] 2025-07-17T09:17:08.6364030Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0032s] [ 98%] 2025-07-17T09:17:08.6364456Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 98%] 2025-07-17T09:17:08.6364886Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0037s] [ 98%] 2025-07-17T09:17:08.6365363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0038s] [ 98%] 2025-07-17T09:17:08.6365832Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 98%] 2025-07-17T09:17:08.6366265Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6366808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6367248Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6367691Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6368118Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6368551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 98%] 2025-07-17T09:17:08.6368975Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0037s] [ 98%] 2025-07-17T09:17:08.6369406Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0037s] [ 98%] 2025-07-17T09:17:08.6369834Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0041s] [ 98%] 2025-07-17T09:17:08.6370278Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0040s] [ 98%] 2025-07-17T09:17:08.6370706Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0040s] [ 98%] 2025-07-17T09:17:08.6371134Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0040s] [ 98%] 2025-07-17T09:17:08.6371560Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0042s] [ 98%] 2025-07-17T09:17:08.6371989Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 98%] 2025-07-17T09:17:08.6372476Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 98%] 2025-07-17T09:17:08.6372954Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6373382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0037s] [ 98%] 2025-07-17T09:17:08.6373913Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0037s] [ 98%] 2025-07-17T09:17:08.6374339Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6374766Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0038s] [ 98%] 2025-07-17T09:17:08.6375195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0044s] [ 98%] 2025-07-17T09:17:08.6375633Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0044s] [ 98%] 2025-07-17T09:17:08.6376050Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0045s] [ 98%] 2025-07-17T09:17:08.6376483Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0044s] [ 98%] 2025-07-17T09:17:08.6376899Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0044s] [ 98%] 2025-07-17T09:17:08.6377331Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_2048_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0044s] [ 98%] 2025-07-17T09:17:08.6377761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0030s] [ 98%] 2025-07-17T09:17:08.6378187Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 98%] 2025-07-17T09:17:08.6378610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 98%] 2025-07-17T09:17:08.6379032Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 98%] 2025-07-17T09:17:08.6379509Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 98%] 2025-07-17T09:17:08.6379974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 98%] 2025-07-17T09:17:08.6380412Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 98%] 2025-07-17T09:17:08.6380945Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 98%] 2025-07-17T09:17:08.6381370Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 98%] 2025-07-17T09:17:08.6381797Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 98%] 2025-07-17T09:17:08.6382212Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 98%] 2025-07-17T09:17:08.6382642Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 98%] 2025-07-17T09:17:08.6383062Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0031s] [ 98%] 2025-07-17T09:17:08.6383491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0033s] [ 98%] 2025-07-17T09:17:08.6383906Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 98%] 2025-07-17T09:17:08.6384328Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0032s] [ 98%] 2025-07-17T09:17:08.6384761Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 98%] 2025-07-17T09:17:08.6385179Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 98%] 2025-07-17T09:17:08.6385604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0037s] [ 98%] 2025-07-17T09:17:08.6386034Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6386454Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0036s] [ 98%] 2025-07-17T09:17:08.6386926Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0038s] [ 98%] 2025-07-17T09:17:08.6387416Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0035s] [ 98%] 2025-07-17T09:17:08.6387949Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_128_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0035s] [ 98%] 2025-07-17T09:17:08.6388380Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0030s] [ 98%] 2025-07-17T09:17:08.6388812Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6389230Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6389655Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0029s] [ 99%] 2025-07-17T09:17:08.6390072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0028s] [ 99%] 2025-07-17T09:17:08.6390496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0029s] [ 99%] 2025-07-17T09:17:08.6390919Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6391355Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6391774Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6392203Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6392623Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 99%] 2025-07-17T09:17:08.6393044Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 99%] 2025-07-17T09:17:08.6393470Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0032s] [ 99%] 2025-07-17T09:17:08.6394060Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 99%] 2025-07-17T09:17:08.6394547Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 99%] 2025-07-17T09:17:08.6395030Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 99%] 2025-07-17T09:17:08.6395508Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6395934Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6396362Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6396792Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6397210Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6397635Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6398058Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6398491Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_16_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0042s] [ 99%] 2025-07-17T09:17:08.6398913Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6399343Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 99%] 2025-07-17T09:17:08.6399770Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6400195Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6400616Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0029s] [ 99%] 2025-07-17T09:17:08.6401089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 99%] 2025-07-17T09:17:08.6401574Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6402003Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6402531Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6402957Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6403382Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 99%] 2025-07-17T09:17:08.6403808Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 99%] 2025-07-17T09:17:08.6404226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0031s] [ 99%] 2025-07-17T09:17:08.6404655Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 99%] 2025-07-17T09:17:08.6405072Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 99%] 2025-07-17T09:17:08.6405496Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 99%] 2025-07-17T09:17:08.6405908Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0031s] [ 99%] 2025-07-17T09:17:08.6406330Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6406756Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 99%] 2025-07-17T09:17:08.6407185Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6407610Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6408089Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6408555Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6408973Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_8_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6409515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6409949Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0029s] [ 99%] 2025-07-17T09:17:08.6410371Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6410802Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6411226Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0028s] [ 99%] 2025-07-17T09:17:08.6411655Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0028s] [ 99%] 2025-07-17T09:17:08.6412079Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6412513Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6412931Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6413363Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6413779Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0033s] [ 99%] 2025-07-17T09:17:08.6414204Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_False_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0033s] [ 99%] 2025-07-17T09:17:08.6414631Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale0_cuda_bfloat16 PASSED [0.0031s] [ 99%] 2025-07-17T09:17:08.6415141Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0031s] [ 99%] 2025-07-17T09:17:08.6415611Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale0_cuda_float16 PASSED [0.0031s] [ 99%] 2025-07-17T09:17:08.6416027Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float16_scale_l1_cuda_float16 PASSED [0.0031s] [ 99%] 2025-07-17T09:17:08.6416551Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale0_cuda_float32 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6416974Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_0_float32_scale_l1_cuda_float32 PASSED [0.0030s] [ 99%] 2025-07-17T09:17:08.6417405Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale0_cuda_bfloat16 PASSED [0.0036s] [ 99%] 2025-07-17T09:17:08.6417835Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_bfloat16_scale_l1_cuda_bfloat16 PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6418257Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale0_cuda_float16 PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6418679Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float16_scale_l1_cuda_float16 PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6419091Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale0_cuda_float32 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6419515Z test_transformers.py::TestSDPACudaOnlyCUDA::test_mem_efficient_attention_vs_math_ref_grads_batch_size_8_seq_len_q_8_seq_len_k_8_head_dim_96_is_causal_True_dropout_p_0_22_float32_scale_l1_cuda_float32 PASSED [0.0034s] [ 99%] 2025-07-17T09:17:08.6419924Z test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_cudnn_nested_type_nested_is_contiguous_True_cuda SKIPPED [0.0001s] (Fused SDPA was not built for this system) [ 99%] 2025-07-17T09:17:08.6420272Z test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_dense_fused_kernel0_cuda PASSED [0.0022s] [ 99%] 2025-07-17T09:17:08.6420602Z test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_dense_fused_kernel1_cuda PASSED [0.0017s] [ 99%] 2025-07-17T09:17:08.6420943Z test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_nested_fused_kernel0_cuda PASSED [0.0069s] [ 99%] 2025-07-17T09:17:08.6421275Z test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_accuracy_type_nested_fused_kernel1_cuda PASSED [0.0065s] [ 99%] 2025-07-17T09:17:08.6421604Z test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_type_dense_is_contiguous_False_cuda PASSED [0.0026s] [ 99%] 2025-07-17T09:17:08.6421934Z test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_type_dense_is_contiguous_True_cuda PASSED [0.0014s] [ 99%] 2025-07-17T09:17:08.6422317Z test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_type_nested_is_contiguous_False_cuda PASSED [0.0175s] [ 99%] 2025-07-17T09:17:08.6422693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_scaled_dot_product_attention_fused_kernels_packed_type_nested_is_contiguous_True_cuda PASSED [0.0175s] [ 99%] 2025-07-17T09:17:08.6423079Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_choice_with_determinism_warn_only_False_cuda SKIPPED [0.0008s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 99%] 2025-07-17T09:17:08.6423448Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_choice_with_determinism_warn_only_True_cuda SKIPPED [0.0007s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 99%] 2025-07-17T09:17:08.6423917Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_False_bfloat16_cuda_bfloat16 PASSED [0.0024s] [ 99%] 2025-07-17T09:17:08.6424277Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_False_float16_cuda_float16 PASSED [0.0019s] [ 99%] 2025-07-17T09:17:08.6424629Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_True_bfloat16_cuda_bfloat16 PASSED [0.0020s] [ 99%] 2025-07-17T09:17:08.6424980Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_False_is_causal_True_float16_cuda_float16 PASSED [0.0020s] [ 99%] 2025-07-17T09:17:08.6425329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_False_bfloat16_cuda_bfloat16 PASSED [0.0019s] [ 99%] 2025-07-17T09:17:08.6425683Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_False_float16_cuda_float16 PASSED [0.0019s] [ 99%] 2025-07-17T09:17:08.6426025Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_True_bfloat16_cuda_bfloat16 PASSED [0.0020s] [ 99%] 2025-07-17T09:17:08.6426376Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_flash_attention_grad_against_math_contiguous_inputs_True_is_causal_True_float16_cuda_float16 PASSED [0.0019s] [ 99%] 2025-07-17T09:17:08.6426693Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_False_is_causal_False_cuda PASSED [0.0019s] [ 99%] 2025-07-17T09:17:08.6427012Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_False_is_causal_True_cuda PASSED [0.0019s] [ 99%] 2025-07-17T09:17:08.6427329Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_True_is_causal_False_cuda PASSED [0.0019s] [ 99%] 2025-07-17T09:17:08.6427637Z test_transformers.py::TestSDPACudaOnlyCUDA::test_sdp_mem_efficient_grad_against_math_contiguous_inputs_True_is_causal_True_cuda PASSED [0.0019s] [ 99%] 2025-07-17T09:17:08.6427864Z test_transformers.py::TestSDPACudaOnlyCUDA::test_singelton_head_dim_stride_ne_1_cuda PASSED [0.0011s] [ 99%] 2025-07-17T09:17:08.6428096Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_1_shape0_cuda PASSED [0.0044s] [ 99%] 2025-07-17T09:17:08.6428323Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_1_shape1_cuda PASSED [0.0120s] [ 99%] 2025-07-17T09:17:08.6428541Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_1_shape2_cuda PASSED [0.0039s] [ 99%] 2025-07-17T09:17:08.6428763Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_1_shape3_cuda PASSED [0.0035s] [ 99%] 2025-07-17T09:17:08.6429088Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_2_shape0_cuda SKIPPED [0.0008s] (No support for LOWER_RIGHT variant for now) [ 99%] 2025-07-17T09:17:08.6429406Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_2_shape1_cuda SKIPPED [0.0007s] (No support for LOWER_RIGHT variant for now) [ 99%] 2025-07-17T09:17:08.6429772Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_2_shape2_cuda SKIPPED [0.0007s] (No support for LOWER_RIGHT variant for now) [ 99%] 2025-07-17T09:17:08.6430141Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_causal_variant_2_shape3_cuda SKIPPED [0.0007s] (No support for LOWER_RIGHT variant for now) [ 99%] 2025-07-17T09:17:08.6430399Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_1_shape0_cuda PASSED [0.4570s] [ 99%] 2025-07-17T09:17:08.6430705Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_1_shape1_cuda PASSED [0.0807s] [ 99%] 2025-07-17T09:17:08.6431004Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_1_shape2_cuda PASSED [0.0786s] [ 99%] 2025-07-17T09:17:08.6431255Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_1_shape3_cuda PASSED [0.0770s] [ 99%] 2025-07-17T09:17:08.6431593Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_2_shape0_cuda SKIPPED [0.0011s] (No support for LOWER_RIGHT variant for now) [ 99%] 2025-07-17T09:17:08.6431929Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_2_shape1_cuda SKIPPED [0.0007s] (No support for LOWER_RIGHT variant for now) [ 99%] 2025-07-17T09:17:08.6432256Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_2_shape2_cuda SKIPPED [0.0007s] (No support for LOWER_RIGHT variant for now) [ 99%] 2025-07-17T09:17:08.6432592Z test_transformers.py::TestAttnBiasCUDA::test_causal_variants_compile_causal_variant_2_shape3_cuda SKIPPED [0.0007s] (No support for LOWER_RIGHT variant for now) [ 99%] 2025-07-17T09:17:08.6432789Z test_transformers.py::TestAttnBiasCUDA::test_is_causal_and_mask_fails_cuda PASSED [0.0012s] [ 99%] 2025-07-17T09:17:08.6433008Z test_transformers.py::TestAttnBiasCUDA::test_is_causal_equals_upper_left_shape0_cuda PASSED [0.0014s] [ 99%] 2025-07-17T09:17:08.6433217Z test_transformers.py::TestAttnBiasCUDA::test_is_causal_equals_upper_left_shape1_cuda PASSED [0.0011s] [ 99%] 2025-07-17T09:17:08.6433432Z test_transformers.py::TestAttnBiasCUDA::test_is_causal_equals_upper_left_shape2_cuda PASSED [0.0011s] [ 99%] 2025-07-17T09:17:08.6433636Z test_transformers.py::TestAttnBiasCUDA::test_is_causal_equals_upper_left_shape3_cuda PASSED [0.0010s] [100%] 2025-07-17T09:17:08.6433641Z 2025-07-17T09:17:08.6434063Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/test_transformers/test_transformers-929fa49726c73d6e.xml - 2025-07-17T09:17:08.6434202Z ======== 8262 passed, 3332 skipped, 648 deselected in 406.02s (0:06:46) ======== 2025-07-17T09:17:08.6434832Z The following tests failed and then succeeded when run in a new process['test/test_transformers.py::TestSDPACudaOnlyCUDA::test_flash_attention_vs_math_ref_grads_batch_size_1_seq_len_q_143_seq_len_k_2048_head_dim_8_is_causal_False_dropout_p_0_22_float16_scale0_enable_gqa_False_n_heads0_cuda_float16'] 2025-07-17T09:17:08.6434836Z 2025-07-17T09:17:08.6435079Z FINISHED PRINTING LOG FILE of test_transformers 1/1 (test/test-reports/test_transformers_1.1_228e957a862f69ea_.log) 2025-07-17T09:17:08.6435083Z 2025-07-17T09:17:08.6435188Z Running test_decomp 10/12 ... [2025-07-17 09:17:06.824534] 2025-07-17T09:17:08.6435275Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:17:08.6435760Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_decomp.py', '-m', 'not serial', '--shard-id=10', '--num-shards=12', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:17:06.824871] 2025-07-17T09:25:37.5491302Z 2025-07-17T09:25:37.5491960Z test_decomp 10/12 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_10.12_8a1f019033861b44_.log 2025-07-17T09:25:37.5632807Z Running 778 items in this shard: test/test_decomp.py::TestDecompCUDA::test_comprehensive_H_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_T_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive___getitem___cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive___getitem___cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___getitem___cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive___radd___cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive___radd___cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___radd___cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rdiv___cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmul___cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmul___cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive___ror___cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rsub___cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rsub___cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rsub___cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive__chunk_cat_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive__segment_reduce_lengths_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive__segment_reduce_offsets_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive__unsafe_masked_index_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive__unsafe_masked_index_put_accumulate_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive__upsample_bilinear2d_aa_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive__upsample_bilinear2d_aa_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_add_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addbmm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addcdiv_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addcdiv_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addcmul_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_alias_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_all_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_amin_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_aminmax_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_angle_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argmax_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argmin_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argmin_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argwhere_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argwhere_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_partial_views_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_partial_views_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_partial_views_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asin_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asinh_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asinh_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atan2_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atan2_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atan2_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atan2_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atanh_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_baddbmm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bernoulli_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bfloat16_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bfloat16_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bfloat16_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bincount_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bitwise_and_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bitwise_or_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_block_diag_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bool_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bool_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bool_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bool_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_broadcast_tensors_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_broadcast_to_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bucketize_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_byte_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cartesian_prod_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cdouble_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cdouble_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ceil_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cfloat_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cfloat_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cfloat_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_char_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_max_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_max_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_min_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_column_stack_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_column_stack_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_combinations_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_conj_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_conj_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_conj_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_contiguous_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_contiguous_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_contiguous_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_copysign_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_corrcoef_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cos_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cos_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cosh_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cosh_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_count_nonzero_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cross_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cross_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummax_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummax_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cumsum_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_deg2rad_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diag_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagflat_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagonal_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagonal_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diff_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diff_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_div_floor_rounding_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_div_floor_rounding_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_div_trunc_rounding_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_div_trunc_rounding_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_double_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_double_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_dsplit_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_like_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_strided_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_strided_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_eq_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_erfinv_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_erfinv_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_exp_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_as_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expm1_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_exponential_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_eye_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fft_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fft_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fft_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fftn_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fftshift_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fftshift_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_hfft2_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_hfft_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_hfft_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_hfftn_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifft_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifftn_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifftn_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifftshift_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifftshift_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfft2_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfft_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_irfft_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfft_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfft_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfft_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfftn_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fliplr_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_power_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_floor_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fmin_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_like_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_like_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_like_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gather_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gather_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gcd_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_geometric_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_geometric_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_geometric_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_geqrf_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gradient_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gradient_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_half_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_heaviside_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_heaviside_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hsplit_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hstack_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hypot_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hypot_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_i0_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_i0_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_i0_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_igammac_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_imag_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_imag_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_add_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_add_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_copy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_copy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_put_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_put_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_reduce_amin_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_reduce_mean_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_reduce_prod_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_select_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_int_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_int_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_int_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isclose_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isfinite_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isfinite_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isinf_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isinf_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isnan_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_istft_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_4inputs_with_extra_args_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_4inputs_with_extra_args_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_binary_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_unary_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_kthvalue_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lcm_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_le_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_le_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lerp_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lgamma_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lgamma_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lgamma_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cholesky_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cond_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cond_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cond_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_eig_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_eig_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_eigh_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_eigvals_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_eigvalsh_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_power_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_norm_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_qr_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_vander_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_vecdot_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_vecdot_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log10_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log1p_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log2_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log2_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log_softmax_with_dtype_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_and_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_and_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_not_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_not_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_xor_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_xor_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_xor_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logit_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logspace_tensor_overload_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logsumexp_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logsumexp_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_long_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mT_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_amax_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_argmax_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_argmin_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_argmin_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_cumprod_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_cumsum_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_cumsum_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_fill_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_norm_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_prod_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_prod_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_select_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_sum_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_var_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_binary_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_binary_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_reduction_no_dim_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_reduction_with_dim_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_reduction_with_dim_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_maximum_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mean_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_median_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_meshgrid_list_of_tensors_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_meshgrid_variadic_tensors_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_meshgrid_variadic_tensors_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_reduction_no_dim_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mm_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mm_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_msort_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mul_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_multinomial_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_multinomial_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_1_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_3_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nan_to_num_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nanmean_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nansum_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nansum_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_narrow_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ne_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_neg_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_empty_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_full_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_ones_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_adaptive_avg_pool3d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_adaptive_max_pool2d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_avg_pool1d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_avg_pool1d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_avg_pool2d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_binary_cross_entropy_with_logits_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_conv3d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_conv_transpose1d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_cosine_embedding_loss_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_dropout3d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_dropout3d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_embedding_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_feature_alpha_dropout_with_train_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_glu_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_hinge_embedding_loss_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_hinge_embedding_loss_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_bicubic_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_bicubic_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_nearest-exact_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_kl_div_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_layer_norm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_logsigmoid_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_unpool1d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_unpool2d_grad_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_unpool2d_grad_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_mish_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_multi_margin_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_one_hot_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_circular_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_circular_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_constant_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_replicate_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_replicate_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pairwise_distance_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_unshuffle_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_unshuffle_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_relu6_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_rms_norm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softmin_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softmin_with_dtype_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_tanhshrink_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nonzero_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_norm_inf_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_normal_in_place_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_normal_number_mean_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ones_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ones_like_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ones_like_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ormqr_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_1_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_3_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_3_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_4_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_positive_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_put_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_put_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_qr_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rad2deg_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randint_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randint_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randint_like_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randn_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randn_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ravel_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_real_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reciprocal_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_renorm_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_repeat_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reshape_as_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reshape_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resize__cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resize_as__cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_conj_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_conj_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_neg_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_neg_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_neg_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rot90_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rot90_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_round_decimals_0_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rsqrt_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rsub_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_add_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_add_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_amax_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_amax_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_amin_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_mean_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_mean_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_mean_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_prod_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_sum_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_searchsorted_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_searchsorted_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sgn_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_short_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signal_windows_gaussian_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signal_windows_general_hamming_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signbit_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sin_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sin_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinc_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinc_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinh_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_slice_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_softmax_with_dtype_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_softmax_with_dtype_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sparse_sampled_addmm_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sparse_sampled_addmm_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_j0_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_j0_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_y0_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_y1_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_chebyshev_polynomial_w_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_entr_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_hermite_polynomial_h_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_hermite_polynomial_he_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i0e_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i0e_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i0e_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i1e_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i1e_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_laguerre_polynomial_l_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_legendre_polynomial_p_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_legendre_polynomial_p_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_modified_bessel_i0_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_v_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_v_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_w_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_spherical_bessel_j0_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_spherical_bessel_j0_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_spherical_bessel_j0_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_zeta_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_zeta_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_zeta_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_list_args_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_with_sizes_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_with_sizes_copy_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_with_sizes_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_with_sizes_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_with_sizes_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sqrt_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sqrt_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_multiple_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_stack_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_std_unbiased_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sub_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_to_size_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_svd_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_svd_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_t_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_t_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_t_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_take_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_take_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tan_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tanh_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tanh_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tensor_split_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tensor_split_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tensor_split_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tile_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tile_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_to_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_to_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_to_sparse_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_topk_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_topk_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapz_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_triu_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_triu_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_true_divide_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_true_divide_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trunc_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unflatten_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unfold_copy_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unique_consecutive_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unique_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsafe_chunk_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsafe_chunk_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsafe_chunk_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsqueeze_copy_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsqueeze_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsqueeze_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsqueeze_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_mean_unbiased_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_as_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_as_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vsplit_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vsplit_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vstack_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vstack_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_xlogy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_xlogy_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zero__cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zeros_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick__batch_norm_with_update_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick__chunk_cat_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick__unsafe_masked_index_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick__unsafe_masked_index_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick__unsafe_masked_index_put_accumulate_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_abs_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_abs_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_acos_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_acosh_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_acosh_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_addcmul_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_addmm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_addmv_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_addr_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_amax_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_amax_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_amax_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_aminmax_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_aminmax_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_aminmax_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_aminmax_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_as_strided_copy_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_as_strided_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_as_strided_scatter_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_asin_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_asin_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_asin_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_asinh_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_asinh_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_atan2_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_atan2_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_atan2_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_atan_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_atan_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_left_shift_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_not_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_not_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_right_shift_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_xor_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_block_diag_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_block_diag_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_block_diag_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_block_diag_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_bucketize_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_cat_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_cat_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_ceil_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_ceil_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_ceil_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_max_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_max_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_max_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_min_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_clone_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_conj_physical_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_conj_physical_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_copysign_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_copysign_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_deg2rad_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_masked_fill_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_mvlgamma_mvlgamma_p_1_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_nn_functional_max_unpool2d_grad_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_split_with_sizes_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_trace_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_count_nonzero_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_cumsum_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_diag_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_diag_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_diag_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_diag_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_diag_embed_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_scatter_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_digamma_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_div_no_rounding_mode_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_div_trunc_rounding_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_empty_like_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_empty_strided_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_empty_strided_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_empty_strided_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_empty_strided_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_eq_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_eq_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_eq_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_erf_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_erfc_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_erfc_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_erfinv_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_erfinv_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_exp_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_expand_copy_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_expand_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_expand_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_expm1_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_expm1_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_expm1_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_exponential_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fft2_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fft2_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fftn_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_hfft2_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_fft_hfft2_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ifftn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ifftn_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ihfft2_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_fft_irfft_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_irfft_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_rfft2_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_rfft_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_rfftn_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fill_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_fill_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_fill_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_fill_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fill_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_flip_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_floor_divide_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fmax_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_fmax_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fmax_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_fmin_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_frexp_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_frexp_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_full_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_ge_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_gt_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_igamma_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_index_add_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_index_copy_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_index_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_index_copy_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_index_fill_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_index_select_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_isinf_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_isnan_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_isnan_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_isneginf_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_isneginf_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_lcm_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_le_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_le_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_le_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_lerp_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_lerp_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_lerp_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_cross_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_linspace_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_linspace_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_linspace_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_log1p_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_log1p_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_log_normal_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_logaddexp_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_logical_and_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_logical_and_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_logical_not_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_logical_xor_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_logit_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_masked_fill_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_mean_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_meshgrid_list_of_tensors_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_meshgrid_variadic_tensors_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_mul_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_mul_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_mvlgamma_mvlgamma_p_3_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_nansum_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_narrow_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_ne_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_neg_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_strided_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_strided_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_strided_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_new_full_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_new_full_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_new_ones_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_new_zeros_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_binary_cross_entropy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_embedding_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_hardshrink_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_hardsigmoid_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_leaky_relu_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_max_unpool2d_grad_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_max_unpool2d_grad_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_relu6_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_norm_fro_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_norm_inf_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_normal_in_place_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_ones_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_permute_copy_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_pow_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_prod_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_randn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_renorm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_roll_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_rot90_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_round_decimals_neg_3_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_rsqrt_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_rsqrt_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_select_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_select_scatter_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_select_scatter_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_sigmoid_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_sigmoid_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_signbit_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_sin_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_sinc_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_sinc_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_sinc_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_sinc_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_sinh_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_slice_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_slice_scatter_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_special_entr_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_special_erfcx_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_special_i0e_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_special_i0e_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_special_i1_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_special_i1e_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_special_log_ndtr_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_special_ndtr_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_special_ndtr_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_special_ndtr_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_special_ndtri_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_special_xlog1py_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_split_list_args_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_sqrt_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_copy_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_multiple_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_stack_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_stack_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_std_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_std_unbiased_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_std_unbiased_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_std_unbiased_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_sum_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_tan_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_tanh_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_trace_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_transpose_copy_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_transpose_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_transpose_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_transpose_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_triu_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_triu_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_unbind_copy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_unbind_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_unbind_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_uniform_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_unsafe_split_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_unsqueeze_copy_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_unsqueeze_copy_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_var_mean_unbiased_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_var_unbiased_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_vdot_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_view_copy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_view_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_view_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_where_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_where_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_xlogy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_zero__cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_zeros_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_zeros_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_zeros_like_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_zeros_like_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_rnn_decomp_module_nn_LSTM_train_mode_cuda_float32 2025-07-17T09:25:37.5772135Z 2025-07-17T09:25:37.5772262Z Running test_decomp 12/12 ... [2025-07-17 09:25:37.549630] 2025-07-17T09:25:37.5772514Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:25:37.5773156Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_decomp.py', '-m', 'not serial', '--shard-id=12', '--num-shards=12', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:25:37.549928] 2025-07-17T09:27:04.0872800Z 2025-07-17T09:27:04.0876453Z test_decomp 2/12 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_2.12_8bb74e22a3aa3900_.log 2025-07-17T09:27:04.1013458Z Running 769 items in this shard: test/test_decomp.py::TestDecompCUDA::test_arange_graph_cuda, test/test_decomp.py::TestDecompCUDA::test_bernoulli_default_cuda, test/test_decomp.py::TestDecompCUDA::test_comprehensive_H_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_H_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_H_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_H_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_H_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_T_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_T_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_T_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_T_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive___radd___cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmatmul___cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmul___cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmul___cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rpow___cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rpow___cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rxor___cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_abs_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acos_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acos_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acos_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acosh_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addcdiv_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addcmul_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addmm_decomposed_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addmv_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addr_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_all_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_allclose_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_allclose_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_amax_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_amin_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_aminmax_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_arange_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_arange_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_arange_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_arange_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argmin_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argsort_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argwhere_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_partial_views_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_scatter_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asin_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asin_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asin_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atan_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atanh_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atanh_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_2d_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_3d_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_baddbmm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bfloat16_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bfloat16_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bitwise_not_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bitwise_or_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bool_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_broadcast_tensors_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_broadcast_to_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bucketize_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_byte_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_byte_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_byte_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cartesian_prod_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cartesian_prod_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cauchy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cdouble_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ceil_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cfloat_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cfloat_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cfloat_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_chalf_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_chalf_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_char_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cholesky_solve_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cholesky_solve_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_chunk_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_chunk_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_min_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_min_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clone_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clone_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_column_stack_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_column_stack_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_conj_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_conj_physical_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_conj_physical_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_constant_pad_nd_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_corrcoef_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cos_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cos_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cosh_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cross_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cross_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummin_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummin_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummin_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummin_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cumprod_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cumsum_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cumulative_trapezoid_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cumulative_trapezoid_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diag_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diag_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diag_embed_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diag_embed_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diag_embed_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagflat_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagonal_scatter_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diff_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_digamma_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_digamma_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_digamma_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_digamma_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_dist_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_div_floor_rounding_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_dot_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_dsplit_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_like_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_permuted_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_permuted_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_permuted_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_permuted_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_strided_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_erf_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_erfinv_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expm1_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expm1_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_eye_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_eye_cuda_float8_e4m3fn, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fft_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fftn_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fftn_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fftshift_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fftshift_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_hfft2_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_hfftn_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifft2_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifft2_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifft_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifftn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfft2_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfft_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfftn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfftn_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfft2_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfft2_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfft_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfftn_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfftn_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fill_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fill_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_flatten_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fliplr_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_flipud_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_power_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_floor_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_floor_divide_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_floor_divide_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_floor_divide_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fmod_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fmod_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gather_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gather_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gradient_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gradient_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gradient_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_histc_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_histc_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hsplit_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hstack_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_i0_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_add_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_put_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_reduce_amax_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_reduce_amin_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_reduce_amin_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_reduce_prod_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_int_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_int_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isclose_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isfinite_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isin_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isin_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isinf_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isinf_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isnan_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isnan_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isneginf_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isreal_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_item_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_binary_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_binary_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_binary_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_binary_return_by_ref_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_kron_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_kron_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_kron_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ldexp_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_le_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cond_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cross_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_diagonal_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_diagonal_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_diagonal_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_eigh_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_ldl_factor_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_lstsq_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_lu_solve_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_lu_solve_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_norm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_norm_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_norm_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_rank_hermitian_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_pinv_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_pinv_singular_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_qr_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_slogdet_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_svd_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_svd_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linspace_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linspace_tensor_overload_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linspace_tensor_overload_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log10_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log1p_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log1p_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log1p_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log_normal_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log_softmax_with_dtype_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logcumsumexp_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_not_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_not_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_or_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_or_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_or_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logit_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logit_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_long_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mT_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_argmin_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_cumprod_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_cumsum_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_cumsum_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_cumsum_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_fill_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_norm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_prod_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_prod_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_select_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_softmin_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_std_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_sum_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_sum_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_var_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_var_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_matmul_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_matmul_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_binary_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_pool2d_with_indices_backward_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_max_reduction_with_dim_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_maximum_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mean_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_meshgrid_list_of_tensors_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_binary_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_reduction_no_dim_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_reduction_with_dim_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_movedim_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_movedim_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_msort_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_3_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_5_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_5_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nanmedian_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nanquantile_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_narrow_copy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_narrow_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_narrow_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_native_dropout_backward_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ne_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ne_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_neg_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_empty_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_empty_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_empty_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_empty_strided_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_adaptive_avg_pool1d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_adaptive_max_pool3d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_avg_pool1d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_channel_shuffle_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_channel_shuffle_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_conv3d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_conv_transpose1d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_conv_transpose2d_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_conv_transpose3d_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_cosine_embedding_loss_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_cosine_similarity_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_ctc_loss_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_elu_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_embedding_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_feature_alpha_dropout_with_train_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_fractional_max_pool2d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_fractional_max_pool3d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_group_norm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_hardtanh_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_area_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_nearest-exact_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_layer_norm_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_leaky_relu_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_margin_ranking_loss_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_margin_ranking_loss_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_pool1d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_pool2d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_unpool1d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_unpool2d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_unpool2d_grad_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_multilabel_margin_loss_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_circular_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_replicate_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pairwise_distance_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_unshuffle_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_poisson_nll_loss_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_poisson_nll_loss_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_prelu_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_relu_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_selu_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_selu_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_silu_complex_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_smooth_l1_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softmin_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softmin_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softplus_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softshrink_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_softsign_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_threshold_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_triplet_margin_loss_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_triplet_margin_loss_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_unfold_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_upsample_nearest_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nonzero_static_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nonzero_static_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_normal_in_place_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ones_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ones_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ones_like_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ones_like_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_copy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_copy_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_pinverse_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polar_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_0_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_3_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_4_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_positive_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_put_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_qr_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rad2deg_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rand_like_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randint_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randint_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randint_like_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randn_like_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ravel_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_real_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_remainder_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_remainder_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_repeat_interleave_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reshape_as_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reshape_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resize_as__cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resize_as__cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_conj_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_neg_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_roll_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_roll_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rot90_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rsub_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scalar_tensor_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scalar_tensor_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scalar_tensor_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_amax_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_amax_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_amin_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_prod_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_prod_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_select_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_select_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_select_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_select_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sgn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sgn_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_short_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_short_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sigmoid_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sigmoid_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sign_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sign_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signal_windows_bartlett_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signal_windows_general_hamming_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signbit_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sin_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinc_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinh_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinh_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_slice_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_slice_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_softmax_with_dtype_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sparse_sampled_addmm_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_j0_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_j0_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_j0_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_j1_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_j1_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_y0_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_chebyshev_polynomial_u_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_chebyshev_polynomial_v_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_chebyshev_polynomial_w_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_erfcx_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_erfcx_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i0e_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i0e_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i1_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i1e_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_legendre_polynomial_p_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_modified_bessel_k1_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_ndtr_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_ndtr_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_ndtri_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_polygamma_special_polygamma_n_0_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_polygamma_special_polygamma_n_0_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_scaled_modified_bessel_k1_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_v_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_w_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_xlog1py_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_list_args_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_with_sizes_copy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sqrt_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_square_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_copy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_multiple_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_stack_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_std_mean_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_std_mean_unbiased_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sub_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_to_size_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_to_size_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_to_size_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_t_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_t_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_to_sparse_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_to_sparse_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_topk_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_torch_ops_aten__flash_attention_forward_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_torch_ops_aten__safe_softmax_default_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_torch_ops_aten__safe_softmax_default_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapezoid_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tril_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tril_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tril_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_triu_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_triu_indices_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_triu_indices_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_true_divide_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_true_divide_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_true_divide_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trunc_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unbind_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unbind_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unflatten_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unflatten_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unfold_copy_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unfold_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unfold_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unique_consecutive_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unique_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unravel_index_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsafe_chunk_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsafe_split_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsafe_split_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_mean_unbiased_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_mean_unbiased_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_mean_unbiased_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_unbiased_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vsplit_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vsplit_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vstack_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_where_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_xlogy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zeros_like_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zeros_like_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick__chunk_cat_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick__chunk_cat_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick__unsafe_masked_index_put_accumulate_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick__unsafe_masked_index_put_accumulate_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick__upsample_bilinear2d_aa_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_abs_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_abs_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_abs_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_acos_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_acos_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_acosh_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_addcmul_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_addmm_decomposed_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_all_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_any_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_asin_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_atan2_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_atan_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_and_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_or_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_block_diag_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_cat_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_ceil_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_ceil_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_max_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_min_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_clone_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_clone_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_conj_physical_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_constant_pad_nd_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_constant_pad_nd_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_dot_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_logsumexp_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_norm_nuc_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_squeeze_multiple_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_tril_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_cos_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_cos_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_count_nonzero_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_cumsum_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_deg2rad_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_diag_embed_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_scatter_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_digamma_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_dist_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_dist_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_div_no_rounding_mode_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_div_no_rounding_mode_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_empty_like_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_empty_strided_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_erfc_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_exp2_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_exp2_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_expand_copy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_expand_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_expand_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_expand_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_eye_cuda_float8_e4m3fnuz, test/test_decomp.py::TestDecompCUDA::test_quick_eye_cuda_float8_e5m2fnuz, test/test_decomp.py::TestDecompCUDA::test_quick_eye_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fft2_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fft2_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fft_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fftn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fftn_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ifft2_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ifft_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ifftn_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ihfftn_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_fft_irfft2_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_irfft2_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_irfft2_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_irfft2_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_irfft2_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_irfftn_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_rfft2_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_rfftn_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_fill_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_flip_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_floor_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_fmax_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_fmin_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fmin_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_fmod_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_fmod_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_fmod_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_frac_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_frac_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_ge_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_ge_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_gt_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_hypot_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_index_add_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_index_fill_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_isin_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_isinf_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_isnan_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_isposinf_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_isposinf_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_item_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_item_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_item_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_item_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_item_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_le_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_lgamma_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_diagonal_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_diagonal_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_linspace_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_log10_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_log10_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_log1p_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_log2_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_log2_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_log_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_log_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_log_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_logical_and_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_logical_and_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_logical_not_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_logical_not_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_logit_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_logspace_tensor_overload_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_masked_fill_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_masked_fill_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_maximum_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_meshgrid_list_of_tensors_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_meshgrid_list_of_tensors_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_minimum_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_minimum_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_mul_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_mul_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nan_to_num_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nansum_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_narrow_copy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_narrow_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_native_batch_norm_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_native_dropout_backward_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_native_layer_norm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_ne_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_strided_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_new_ones_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_new_zeros_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_new_zeros_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_elu_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_embedding_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_hardsigmoid_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_hardswish_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_hardtanh_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_max_unpool2d_grad_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_max_unpool3d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_max_unpool3d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_pad_constant_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_prelu_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_prelu_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_relu6_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_relu6_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_relu_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_softshrink_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_norm_inf_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_norm_nuc_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_normal_number_mean_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_ones_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_ones_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_ones_like_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_permute_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_prod_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_remainder_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_repeat_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_repeat_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_repeat_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_roll_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_roll_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_rot90_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_round_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_rsqrt_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_rsub_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_select_scatter_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_sigmoid_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_sign_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_sign_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_signbit_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_signbit_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_sin_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_sin_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_sinh_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_slice_scatter_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_softmax_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_special_erfcx_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_special_i0e_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_special_i1_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_special_i1_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_special_i1_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_special_ndtr_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_special_ndtr_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_special_ndtr_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_special_xlog1py_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_special_zeta_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_special_zeta_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_special_zeta_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_split_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_split_list_args_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_split_list_args_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_split_with_sizes_copy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_split_with_sizes_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_split_with_sizes_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_sqrt_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_copy_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_copy_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_multiple_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_multiple_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_std_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_std_mean_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_std_mean_unbiased_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_std_mean_unbiased_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_std_unbiased_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_sub_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_sub_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_sum_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_tanh_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_trace_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_trace_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_transpose_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_transpose_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_tril_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_unbind_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_unbind_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_copy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_unsafe_split_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_unsqueeze_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_unsqueeze_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_var_mean_unbiased_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_vdot_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_vdot_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_view_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_where_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_where_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_xlogy_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_xlogy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_zero__cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_zeros_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_zeros_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_zeros_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_uniform_cuda, test/test_decomp.py::DecompOneOffTestsCUDA::test_native_layer_norm_cpu_decomp_cuda, test/test_decomp.py::HasDecompTest::test_has_decomposition 2025-07-17T09:27:04.1149217Z 2025-07-17T09:27:04.1149334Z Running test_foreach 1/1 ... [2025-07-17 09:27:04.087999] 2025-07-17T09:27:04.1149588Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:27:04.1150399Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_foreach.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:27:04.088300] 2025-07-17T09:31:39.6633554Z 2025-07-17T09:31:39.6637717Z test_foreach 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_foreach_1.1_216454f82a1d9403_.log 2025-07-17T09:31:39.7448021Z Running 3577 items in this shard: test/test_foreach.py::TestForeachCUDA::test_0dim_tensor_overload_cpu_ok_cuda, test/test_foreach.py::TestForeachCUDA::test_0dim_tensor_overload_exception_cuda, test/test_foreach.py::TestForeachCUDA::test_add_scalar_with_empty_list_and_empty_tensor_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_add_scalar_with_empty_list_and_empty_tensor_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_add_scalar_with_empty_list_and_empty_tensor_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_add_scalar_with_empty_list_and_empty_tensor_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_add_scalar_with_empty_list_and_empty_tensor_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_add_scalar_with_empty_list_and_empty_tensor_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_add_scalar_with_empty_list_and_empty_tensor_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_add_scalar_with_empty_list_and_empty_tensor_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_add_scalar_with_empty_list_and_empty_tensor_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_add_scalar_with_empty_list_and_empty_tensor_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_add_scalar_with_empty_list_and_empty_tensor_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_abs_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_acos_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_add_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_addcdiv_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_addcmul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_asin_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_atan_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_ceil_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_clamp_max_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_clamp_min_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_cos_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_cosh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_div_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_erf_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_erfc_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_exp_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_expm1_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_floor_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_frac_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_lerp_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_lgamma_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_log10_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_log1p_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_log2_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_log_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_max_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_maximum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_minimum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_mul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_neg_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_norm_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_pow_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_reciprocal_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_round_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_rsqrt_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_sigmoid_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_sign_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_sin_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_sinh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_sqrt_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_sub_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_tan_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_tanh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_trunc_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_all_zero_size_tensors_do_not_launch_kernel__foreach_zero_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_abs_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_abs_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_abs_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_abs_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_acos_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_acos_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_acos_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_acos_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_add_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_add_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_add_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_add_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_addcdiv_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_addcdiv_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_addcdiv_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_addcdiv_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_addcmul_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_addcmul_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_addcmul_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_addcmul_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_asin_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_asin_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_asin_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_asin_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_atan_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_atan_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_atan_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_atan_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_ceil_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_ceil_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_ceil_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_ceil_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_clamp_max_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_clamp_max_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_clamp_max_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_clamp_max_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_clamp_min_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_clamp_min_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_clamp_min_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_clamp_min_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_copy_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_copy_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_copy_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_copy_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_cos_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_cos_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_cos_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_cos_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_cosh_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_cosh_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_cosh_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_cosh_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_div_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_div_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_div_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_div_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_erf_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_erf_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_erf_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_erf_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_erfc_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_erfc_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_erfc_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_erfc_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_exp_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_exp_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_exp_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_exp_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_expm1_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_expm1_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_expm1_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_expm1_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_floor_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_floor_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_floor_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_floor_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_frac_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_frac_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_frac_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_frac_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_lerp_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_lerp_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_lerp_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_lerp_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_lgamma_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_lgamma_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_lgamma_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_lgamma_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log10_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log10_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log10_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log10_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log1p_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log1p_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log1p_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log1p_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log2_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log2_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log2_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log2_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_log_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_max_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_max_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_max_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_max_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_maximum_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_maximum_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_maximum_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_maximum_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_minimum_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_minimum_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_minimum_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_minimum_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_mul_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_mul_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_mul_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_mul_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_neg_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_neg_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_neg_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_neg_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_norm_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_norm_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_norm_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_norm_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_pow_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_pow_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_pow_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_pow_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_reciprocal_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_reciprocal_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_reciprocal_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_reciprocal_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_round_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_round_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_round_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_round_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_rsqrt_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_rsqrt_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_rsqrt_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_rsqrt_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sigmoid_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sigmoid_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sigmoid_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sigmoid_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sign_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sign_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sign_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sign_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sin_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sin_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sin_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sin_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sinh_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sinh_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sinh_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sinh_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sqrt_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sqrt_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sqrt_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sqrt_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sub_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sub_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sub_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_sub_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_tan_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_tan_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_tan_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_tan_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_tanh_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_tanh_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_tanh_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_tanh_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_trunc_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_trunc_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_trunc_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_trunc_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_zero_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_zero_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_zero_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_autodiff__foreach_zero_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_max_use_cuda_graph_False_w_empty_False_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_max_use_cuda_graph_False_w_empty_False_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_max_use_cuda_graph_False_w_empty_True_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_max_use_cuda_graph_False_w_empty_True_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_max_use_cuda_graph_True_w_empty_False_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_max_use_cuda_graph_True_w_empty_False_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_max_use_cuda_graph_True_w_empty_True_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_max_use_cuda_graph_True_w_empty_True_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_norm_use_cuda_graph_False_w_empty_False_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_norm_use_cuda_graph_False_w_empty_False_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_norm_use_cuda_graph_False_w_empty_True_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_norm_use_cuda_graph_False_w_empty_True_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_norm_use_cuda_graph_True_w_empty_False_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_norm_use_cuda_graph_True_w_empty_False_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_norm_use_cuda_graph_True_w_empty_True_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_big_num_tensors__foreach_norm_use_cuda_graph_True_w_empty_True_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_add_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_add_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_add_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_add_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_clamp_max_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_clamp_max_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_clamp_max_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_clamp_max_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_clamp_min_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_clamp_min_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_clamp_min_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_clamp_min_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_div_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_div_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_div_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_div_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_maximum_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_maximum_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_maximum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_maximum_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_minimum_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_minimum_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_minimum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_minimum_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_mul_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_mul_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_mul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_mul_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_pow_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_pow_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_pow_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_pow_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_sub_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_sub_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_sub_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_float_inf_nan__foreach_sub_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_add_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_add_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_add_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_add_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_add_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_add_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_add_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_add_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_add_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_add_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_add_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_add_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_max_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_max_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_max_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_max_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_max_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_max_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_max_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_max_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_max_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_max_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_max_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_max_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_min_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_min_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_min_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_min_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_min_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_min_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_min_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_min_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_min_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_min_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_min_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_clamp_min_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_div_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_div_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_div_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_div_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_div_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_div_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_div_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_div_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_div_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_div_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_div_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_div_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_maximum_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_maximum_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_maximum_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_maximum_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_maximum_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_maximum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_maximum_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_maximum_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_maximum_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_maximum_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_maximum_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_maximum_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_minimum_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_minimum_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_minimum_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_minimum_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_minimum_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_minimum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_minimum_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_minimum_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_minimum_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_minimum_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_minimum_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_minimum_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_mul_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_mul_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_mul_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_mul_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_mul_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_mul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_mul_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_mul_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_mul_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_mul_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_mul_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_mul_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_pow_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_pow_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_pow_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_pow_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_pow_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_pow_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_pow_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_pow_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_pow_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_pow_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_pow_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_pow_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_sub_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_sub_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_sub_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_sub_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_sub_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_sub_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_sub_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_sub_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_sub_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_sub_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_sub_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_error_cases__foreach_sub_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_add_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_add_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_add_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_add_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_add_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_add_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_add_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_add_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_add_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_add_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_add_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_add_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_max_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_max_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_max_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_max_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_max_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_max_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_max_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_max_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_max_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_max_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_max_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_max_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_min_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_min_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_min_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_min_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_min_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_min_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_min_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_min_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_min_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_min_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_min_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_clamp_min_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_div_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_div_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_div_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_div_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_div_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_div_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_div_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_div_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_div_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_div_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_div_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_div_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_maximum_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_maximum_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_maximum_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_maximum_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_maximum_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_maximum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_maximum_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_maximum_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_maximum_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_maximum_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_maximum_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_maximum_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_minimum_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_minimum_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_minimum_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_minimum_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_minimum_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_minimum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_minimum_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_minimum_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_minimum_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_minimum_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_minimum_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_minimum_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_mul_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_mul_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_mul_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_mul_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_mul_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_mul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_mul_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_mul_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_mul_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_mul_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_mul_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_mul_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_pow_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_pow_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_pow_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_pow_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_pow_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_pow_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_pow_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_pow_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_pow_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_pow_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_pow_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_pow_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_sub_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_sub_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_sub_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_sub_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_sub_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_sub_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_sub_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_sub_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_sub_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_sub_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_sub_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_list_slow_path__foreach_sub_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_different_tensor_dtypes__foreach_add_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_different_tensor_dtypes__foreach_clamp_max_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_different_tensor_dtypes__foreach_clamp_min_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_different_tensor_dtypes__foreach_div_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_different_tensor_dtypes__foreach_maximum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_different_tensor_dtypes__foreach_minimum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_different_tensor_dtypes__foreach_mul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_different_tensor_dtypes__foreach_pow_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_different_tensor_dtypes__foreach_sub_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_add_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_max_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_clamp_min_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_div_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_maximum_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_minimum_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_mul_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_pow_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_scalar_with_overlapping_tensors__foreach_sub_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_add_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_add_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_add_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_add_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_add_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_add_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_add_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_add_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_add_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_add_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_add_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_add_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_max_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_clamp_min_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_div_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_div_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_div_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_div_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_div_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_div_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_div_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_div_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_div_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_div_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_div_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_div_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_maximum_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_maximum_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_maximum_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_maximum_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_maximum_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_maximum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_maximum_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_maximum_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_maximum_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_maximum_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_maximum_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_maximum_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_minimum_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_minimum_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_minimum_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_minimum_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_minimum_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_minimum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_minimum_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_minimum_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_minimum_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_minimum_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_minimum_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_minimum_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_mul_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_mul_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_mul_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_mul_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_mul_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_mul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_mul_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_mul_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_mul_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_mul_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_mul_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_mul_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_pow_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_pow_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_pow_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_pow_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_pow_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_pow_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_pow_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_pow_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_pow_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_pow_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_pow_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_pow_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_sub_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_sub_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_sub_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_sub_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_sub_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_sub_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_sub_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_sub_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_sub_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_sub_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_sub_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_tensors_on_different_devices__foreach_sub_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_False_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_False_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_False_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_False_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_False_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_False_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_False_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_False_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_False_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_False_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_False_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_False_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_True_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_True_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_True_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_True_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_True_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_True_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_True_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_True_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_True_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_binary_op_with_scalar_self_support__foreach_pow_is_fastpath_True_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_div_reciprocal_cuda, test/test_foreach.py::TestForeachCUDA::test_foreach_check_stride_ignore_dims_of_one_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_different_device_inputs__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_different_device_inputs__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_different_device_inputs__foreach_copy_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_different_device_inputs__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_different_device_inputs__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_different_device_inputs__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_different_device_inputs__foreach_copy_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_different_device_inputs__foreach_copy_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_different_device_inputs__foreach_copy_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_different_device_inputs__foreach_copy_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_different_device_inputs__foreach_copy_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_different_device_inputs__foreach_copy_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_device_inputs__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_device_inputs__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_device_inputs__foreach_copy_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_device_inputs__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_device_inputs__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_device_inputs__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_device_inputs__foreach_copy_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_device_inputs__foreach_copy_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_device_inputs__foreach_copy_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_device_inputs__foreach_copy_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_device_inputs__foreach_copy_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_device_inputs__foreach_copy_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes__foreach_copy_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_foreach_copy_with_multi_dtypes_large_input_cuda, test/test_foreach.py::TestForeachCUDA::test_foreach_l2_large_value_input__foreach_norm_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_l2_large_value_input__foreach_norm_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_False_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_False_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_False_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_False_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_False_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_False_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_False_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_False_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_False_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_False_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_False_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_False_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_True_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_True_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_True_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_True_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_True_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_True_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_True_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_True_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_True_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_max_w_empty_True_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_False_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_False_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_False_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_False_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_False_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_False_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_False_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_False_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_False_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_False_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_False_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_False_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_True_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_True_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_True_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_True_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_True_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_True_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_True_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_True_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_True_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_foreach_reduce_large_input__foreach_norm_w_empty_True_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_abs_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_acos_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_add_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_addcdiv_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_addcmul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_asin_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_atan_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_ceil_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_clamp_max_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_clamp_min_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_copy_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_cos_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_cosh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_div_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_erf_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_erfc_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_exp_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_expm1_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_floor_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_frac_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_lerp_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_lgamma_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_log10_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_log1p_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_log2_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_log_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_maximum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_minimum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_mul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_neg_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_pow_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_reciprocal_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_round_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_rsqrt_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_sigmoid_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_sign_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_sin_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_sinh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_sqrt_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_sub_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_tan_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_tanh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_trunc_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_inplace_foreach_leaf_check_and_grad_fn__foreach_zero_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_lifetime_of_grad_fn_when_result_is_saved__foreach_exp_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_lifetime_of_grad_fn_when_result_is_saved__foreach_expm1_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_lifetime_of_grad_fn_when_result_is_saved__foreach_pow_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_lifetime_of_grad_fn_when_result_is_saved__foreach_reciprocal_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_lifetime_of_grad_fn_when_result_is_saved__foreach_rsqrt_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_lifetime_of_grad_fn_when_result_is_saved__foreach_sigmoid_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_lifetime_of_grad_fn_when_result_is_saved__foreach_sqrt_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_lifetime_of_grad_fn_when_result_is_saved__foreach_tan_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_lifetime_of_grad_fn_when_result_is_saved__foreach_tanh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_abs_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_acos_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_add_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_addcdiv_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_addcmul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_asin_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_atan_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_ceil_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_clamp_max_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_clamp_min_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_cos_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_cosh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_div_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_erf_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_erfc_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_exp_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_expm1_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_floor_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_frac_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_lerp_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_lgamma_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_log10_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_log1p_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_log2_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_log_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_maximum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_minimum_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_mul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_neg_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_pow_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_reciprocal_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_round_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_rsqrt_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_sigmoid_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_sign_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_sin_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_sinh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_sqrt_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_sub_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_tan_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_tanh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_outplace_with_invalid_grads__foreach_trunc_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_abs_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_acos_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_add_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcdiv_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_addcmul_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_asin_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_atan_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_ceil_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_max_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_clamp_min_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_copy_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cos_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_cosh_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_div_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erf_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_erfc_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_exp_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_expm1_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_floor_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_frac_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lerp_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_lgamma_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log10_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log1p_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log2_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_log_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_max_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_maximum_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_minimum_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_mul_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_neg_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_norm_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_pow_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_reciprocal_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_round_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_rsqrt_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sigmoid_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sign_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sin_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sinh_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sqrt_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_sub_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tan_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_tanh_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_trunc_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_fastpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_inplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_inplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_inplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_inplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_inplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_inplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_inplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_inplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_inplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_inplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_inplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_inplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_outplace_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_outplace_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_outplace_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_outplace_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_outplace_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_outplace_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_outplace_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_outplace_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_outplace_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_outplace_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_outplace_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_parity__foreach_zero_slowpath_outplace_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_tensors_on_different_devices__foreach_addcdiv_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_tensors_on_different_devices__foreach_addcdiv_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_tensors_on_different_devices__foreach_addcmul_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_tensors_on_different_devices__foreach_addcmul_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_False_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_False_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_False_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_False_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_False_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_False_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_False_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_False_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_False_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_False_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_False_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_False_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcdiv_is_fastpath_True_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_False_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_False_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_False_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_False_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_False_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_False_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_False_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_False_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_False_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_False_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_False_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_False_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_pointwise_op_with_tensor_of_scalarlist_overload__foreach_addcmul_is_fastpath_True_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_tensors_grouping_cuda, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_abs_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_abs_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_abs_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_abs_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_abs_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_abs_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_abs_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_abs_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_abs_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_abs_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_abs_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_abs_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_acos_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_acos_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_acos_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_acos_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_acos_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_acos_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_acos_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_acos_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_acos_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_acos_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_acos_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_acos_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_asin_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_asin_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_asin_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_asin_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_asin_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_asin_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_asin_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_asin_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_asin_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_asin_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_asin_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_asin_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_atan_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_atan_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_atan_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_atan_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_atan_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_atan_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_atan_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_atan_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_atan_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_atan_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_atan_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_atan_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_ceil_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_ceil_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_ceil_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_ceil_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_ceil_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_ceil_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_ceil_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_ceil_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_ceil_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_ceil_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_ceil_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_ceil_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cos_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cos_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cos_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cos_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cos_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cos_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cos_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cos_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cos_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cos_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cos_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cos_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cosh_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cosh_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cosh_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cosh_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cosh_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cosh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cosh_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cosh_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cosh_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cosh_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cosh_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_cosh_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erf_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erf_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erf_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erf_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erf_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erf_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erf_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erf_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erf_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erf_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erf_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erf_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erfc_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erfc_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erfc_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erfc_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erfc_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erfc_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erfc_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erfc_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erfc_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erfc_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erfc_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_erfc_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_exp_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_exp_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_exp_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_exp_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_exp_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_exp_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_exp_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_exp_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_exp_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_exp_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_exp_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_exp_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_expm1_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_expm1_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_expm1_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_expm1_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_expm1_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_expm1_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_expm1_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_expm1_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_expm1_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_expm1_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_expm1_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_expm1_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_floor_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_floor_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_floor_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_floor_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_floor_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_floor_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_floor_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_floor_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_floor_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_floor_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_floor_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_floor_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_frac_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_frac_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_frac_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_frac_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_frac_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_frac_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_frac_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_frac_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_frac_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_frac_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_frac_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_frac_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_lgamma_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_lgamma_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_lgamma_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_lgamma_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_lgamma_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_lgamma_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_lgamma_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_lgamma_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_lgamma_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_lgamma_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_lgamma_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_lgamma_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log10_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log10_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log10_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log10_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log10_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log10_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log10_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log10_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log10_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log10_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log10_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log10_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log1p_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log1p_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log1p_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log1p_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log1p_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log1p_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log1p_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log1p_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log1p_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log1p_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log1p_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log1p_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log2_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log2_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log2_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log2_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log2_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log2_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log2_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log2_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log2_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log2_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log2_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log2_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_log_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_neg_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_neg_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_neg_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_neg_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_neg_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_neg_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_neg_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_neg_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_neg_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_neg_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_neg_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_neg_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_reciprocal_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_round_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_round_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_round_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_round_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_round_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_round_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_round_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_round_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_round_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_round_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_round_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_round_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_rsqrt_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_rsqrt_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_rsqrt_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_rsqrt_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_rsqrt_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_rsqrt_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_rsqrt_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_rsqrt_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_rsqrt_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_rsqrt_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_rsqrt_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_rsqrt_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sigmoid_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sign_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sign_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sign_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sign_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sign_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sign_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sign_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sign_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sign_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sign_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sign_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sign_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sin_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sin_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sin_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sin_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sin_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sin_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sin_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sin_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sin_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sin_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sin_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sin_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sinh_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sinh_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sinh_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sinh_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sinh_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sinh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sinh_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sinh_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sinh_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sinh_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sinh_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sinh_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sqrt_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sqrt_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sqrt_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sqrt_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sqrt_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sqrt_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sqrt_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sqrt_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sqrt_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sqrt_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sqrt_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_sqrt_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tan_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tan_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tan_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tan_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tan_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tan_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tan_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tan_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tan_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tan_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tan_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tan_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tanh_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tanh_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tanh_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tanh_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tanh_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tanh_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tanh_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tanh_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tanh_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tanh_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tanh_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_tanh_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_trunc_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_trunc_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_trunc_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_trunc_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_trunc_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_trunc_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_trunc_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_trunc_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_trunc_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_trunc_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_trunc_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_trunc_cuda_uint8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_zero_cuda_bfloat16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_zero_cuda_bool, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_zero_cuda_complex128, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_zero_cuda_complex64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_zero_cuda_float16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_zero_cuda_float32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_zero_cuda_float64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_zero_cuda_int16, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_zero_cuda_int32, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_zero_cuda_int64, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_zero_cuda_int8, test/test_foreach.py::TestForeachCUDA::test_unary_op_tensors_on_different_devices__foreach_zero_cuda_uint8 2025-07-17T09:31:39.8229162Z 2025-07-17T09:31:39.8229359Z GITHUB_RUN_ID, GITHUB_RUN_ATTEMPT, or ARTIFACTS_FILE_SUFFIX not set, not uploading 2025-07-17T09:31:39.8229755Z Running torch_np/numpy_tests/linalg/test_linalg 1/1 ... [2025-07-17 09:31:39.667275] 2025-07-17T09:31:39.8230065Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:31:39.8230286Z Uploading artifacts took 0.00 seconds 2025-07-17T09:31:39.8230994Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'torch_np/numpy_tests/linalg/test_linalg.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:31:39.667605] 2025-07-17T09:31:47.0466656Z 2025-07-17T09:31:47.0468001Z torch_np/numpy_tests/linalg/test_linalg 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.linalg.test_linalg_1.1_60c6eb2474a20917_.log 2025-07-17T09:31:47.0520725Z Running 268 items in this shard: test/torch_np/numpy_tests/linalg/test_linalg.py::TestSolve::test_0_size, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSolve::test_0_size_k, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSolve::test_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSolve::test_generalized_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSolve::test_generalized_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSolve::test_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSolve::test_types_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSolve::test_types_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSolve::test_types_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSolve::test_types_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestInv::test_0_size, test/torch_np/numpy_tests/linalg/test_linalg.py::TestInv::test_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestInv::test_generalized_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestInv::test_generalized_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestInv::test_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestInv::test_types_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestInv::test_types_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestInv::test_types_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestInv::test_types_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvals::test_0_size, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvals::test_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvals::test_generalized_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvals::test_generalized_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvals::test_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvals::test_types_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvals::test_types_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvals::test_types_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvals::test_types_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEig::test_0_size, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEig::test_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEig::test_generalized_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEig::test_generalized_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEig::test_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEig::test_types_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEig::test_types_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEig::test_types_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEig::test_types_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVD::test_empty_identity, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVD::test_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVD::test_generalized_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVD::test_generalized_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVD::test_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVD::test_types_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVD::test_types_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVD::test_types_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVD::test_types_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVDHermitian::test_empty_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVDHermitian::test_generalized_empty_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVDHermitian::test_generalized_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVDHermitian::test_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVDHermitian::test_types_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVDHermitian::test_types_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVDHermitian::test_types_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestSVDHermitian::test_types_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCond::test_basic_nonsvd, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCond::test_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCond::test_generalized_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCond::test_generalized_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCond::test_nan, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCond::test_singular, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCond::test_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCond::test_stacked_singular, test/torch_np/numpy_tests/linalg/test_linalg.py::TestPinv::test_empty_nonsq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestPinv::test_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestPinv::test_generalized_empty_nonsq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestPinv::test_generalized_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestPinv::test_generalized_nonsq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestPinv::test_generalized_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestPinv::test_nonsq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestPinv::test_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestPinvHermitian::test_empty_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestPinvHermitian::test_generalized_empty_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestPinvHermitian::test_generalized_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestPinvHermitian::test_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestDet::test_0_size, test/torch_np/numpy_tests/linalg/test_linalg.py::TestDet::test_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestDet::test_generalized_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestDet::test_generalized_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestDet::test_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestDet::test_types_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestDet::test_types_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestDet::test_types_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestDet::test_types_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestDet::test_zero, test/torch_np/numpy_tests/linalg/test_linalg.py::TestLstsq::test_empty_a_b_m_0_n_0_n_rhs_0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestLstsq::test_empty_a_b_m_0_n_4_n_rhs_1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestLstsq::test_empty_a_b_m_0_n_4_n_rhs_2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestLstsq::test_empty_a_b_m_4_n_0_n_rhs_1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestLstsq::test_empty_a_b_m_4_n_0_n_rhs_2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestLstsq::test_empty_a_b_m_4_n_2_n_rhs_2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestLstsq::test_empty_nonsq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestLstsq::test_empty_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestLstsq::test_future_rcond, test/torch_np/numpy_tests/linalg/test_linalg.py::TestLstsq::test_incompatible_dims, test/torch_np/numpy_tests/linalg/test_linalg.py::TestLstsq::test_nonsq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestLstsq::test_sq_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvalshCases::test_generalized_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvalshCases::test_generalized_empty_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvalshCases::test_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvalshCases::test_empty_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvalsh::test_0_size, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvalsh::test_UPLO, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvalsh::test_invalid, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvalsh::test_types_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvalsh::test_types_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvalsh::test_types_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigvalsh::test_types_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEighCases::test_generalized_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEighCases::test_generalized_empty_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEighCases::test_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEighCases::test_empty_herm_cases, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigh::test_0_size, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigh::test_UPLO, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigh::test_invalid, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigh::test_types_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigh::test_types_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigh::test_types_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestEigh::test_types_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNorm_NonSystematic::test_intmin, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormDouble::test_axis, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormDouble::test_bad_args, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormDouble::test_empty, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormDouble::test_keepdims, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormDouble::test_matrix_2x2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormDouble::test_matrix_3x3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormDouble::test_matrix_empty, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormDouble::test_matrix_return_type, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormDouble::test_vector, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormDouble::test_vector_return_type, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormSingle::test_axis, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormSingle::test_bad_args, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormSingle::test_empty, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormSingle::test_keepdims, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormSingle::test_matrix_2x2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormSingle::test_matrix_3x3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormSingle::test_matrix_empty, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormSingle::test_matrix_return_type, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormSingle::test_vector, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormSingle::test_vector_return_type, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormInt64::test_axis, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormInt64::test_bad_args, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormInt64::test_empty, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormInt64::test_keepdims, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormInt64::test_matrix_2x2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormInt64::test_matrix_3x3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormInt64::test_matrix_empty, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormInt64::test_matrix_return_type, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormInt64::test_vector, test/torch_np/numpy_tests/linalg/test_linalg.py::TestNormInt64::test_vector_return_type, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMatrixRank::test_matrix_rank, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMatrixRank::test_reduced_rank, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMatrixRank::test_symmetric_rank, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_mode_all_but_economic, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_mode_raw, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_qr_empty_m_0_n_0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_qr_empty_m_0_n_3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_qr_empty_m_3_n_0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size0_outer_size0_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size0_outer_size0_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size0_outer_size0_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size0_outer_size0_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size0_outer_size1_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size0_outer_size1_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size0_outer_size1_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size0_outer_size1_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size0_outer_size2_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size0_outer_size2_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size0_outer_size2_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size0_outer_size2_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size1_outer_size0_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size1_outer_size0_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size1_outer_size0_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size1_outer_size0_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size1_outer_size1_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size1_outer_size1_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size1_outer_size1_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size1_outer_size1_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size1_outer_size2_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size1_outer_size2_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size1_outer_size2_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size1_outer_size2_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size2_outer_size0_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size2_outer_size0_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size2_outer_size0_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size2_outer_size0_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size2_outer_size1_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size2_outer_size1_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size2_outer_size1_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size2_outer_size1_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size2_outer_size2_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size2_outer_size2_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size2_outer_size2_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size2_outer_size2_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size3_outer_size0_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size3_outer_size0_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size3_outer_size0_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size3_outer_size0_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size3_outer_size1_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size3_outer_size1_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size3_outer_size1_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size3_outer_size1_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size3_outer_size2_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size3_outer_size2_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size3_outer_size2_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size3_outer_size2_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size4_outer_size0_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size4_outer_size0_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size4_outer_size0_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size4_outer_size0_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size4_outer_size1_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size4_outer_size1_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size4_outer_size1_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size4_outer_size1_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size4_outer_size2_dt0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size4_outer_size2_dt1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size4_outer_size2_dt2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestQR::test_stacked_inputs_size4_outer_size2_dt3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_0_size, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape0_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape0_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape0_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape0_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape1_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape1_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape1_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape1_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape2_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape2_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape2_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape2_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape3_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape3_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape3_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape3_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape4_dtype0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape4_dtype1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape4_dtype2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestCholesky::test_basic_property_shape4_dtype3, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMisc::test_byteorder_check, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMisc::test_generalized_raise_multiloop, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMisc::test_sdot_bug_8577, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMisc::test_xerbla_override, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMultiDot::test_basic_function_with_dynamic_programming_optimization, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMultiDot::test_basic_function_with_three_arguments, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMultiDot::test_basic_function_with_two_arguments, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMultiDot::test_dynamic_programming_logic, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMultiDot::test_dynamic_programming_optimization_and_out, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMultiDot::test_three_arguments_and_out, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMultiDot::test_too_few_input_arrays, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMultiDot::test_two_arguments_and_out, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMultiDot::test_vector_as_first_and_last_argument, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMultiDot::test_vector_as_first_argument, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMultiDot::test_vector_as_last_argument, test/torch_np/numpy_tests/linalg/test_linalg.py::TestTensorinv::test_non_square_handling_arr0_ind_2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestTensorinv::test_non_square_handling_arr1_ind_1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestTensorinv::test_tensorinv_ind_limit_ind_-2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestTensorinv::test_tensorinv_ind_limit_ind_0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestTensorinv::test_tensorinv_result, test/torch_np/numpy_tests/linalg/test_linalg.py::TestTensorinv::test_tensorinv_shape_shape0_ind_2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestTensorinv::test_tensorinv_shape_shape1_ind_1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestTensorsolve::test_non_square_handling_a0_axes0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestTensorsolve::test_non_square_handling_a1_axes1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestTensorsolve::test_tensorsolve_result_shape0, test/torch_np/numpy_tests/linalg/test_linalg.py::TestTensorsolve::test_tensorsolve_result_shape1, test/torch_np/numpy_tests/linalg/test_linalg.py::TestTensorsolve::test_tensorsolve_result_shape2, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMisc2::test_blas64_dot, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMisc2::test_blas64_geqrf_lwork_smoketest, test/torch_np/numpy_tests/linalg/test_linalg.py::TestMisc2::test_unsupported_commontype 2025-07-17T09:31:47.0572597Z 2025-07-17T09:31:47.0572724Z Running test_jit_fuser_te 1/1 ... [2025-07-17 09:31:47.046810] 2025-07-17T09:31:47.0572995Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:31:47.0573629Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_jit_fuser_te.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:31:47.047111] 2025-07-17T09:35:56.0729034Z 2025-07-17T09:35:56.0730301Z test_decomp 12/12 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_12.12_f87f800a58a293f8_.log 2025-07-17T09:35:56.0860147Z Running 725 items in this shard: test/test_decomp.py::TestDecompCUDA::test_comprehensive_T_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_T_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___getitem___cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive___getitem___cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive___radd___cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive___radd___cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rdiv___cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmatmul___cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmatmul___cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmatmul___cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmod___cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rmul___cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive___rsub___cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive__chunk_cat_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive__chunk_cat_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive__softmax_backward_data_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive__unsafe_masked_index_put_accumulate_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive__unsafe_masked_index_put_accumulate_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive__unsafe_masked_index_put_accumulate_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive__unsafe_masked_index_put_accumulate_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive__upsample_bilinear2d_aa_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_abs_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acos_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acos_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_acos_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_add_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addcmul_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_addcmul_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_alias_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_alias_copy_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_all_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_all_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_allclose_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_amax_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_amin_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_aminmax_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_angle_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_argmax_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_as_strided_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asin_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asin_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asinh_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asinh_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_asinh_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atanh_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_1d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_2d_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_3d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_3d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_atleast_3d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_baddbmm_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bfloat16_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bitwise_left_shift_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bitwise_left_shift_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bitwise_left_shift_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_block_diag_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bmm_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bool_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_broadcast_to_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bucketize_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bucketize_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_bucketize_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cat_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cat_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cat_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cdouble_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cdouble_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cfloat_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_char_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_char_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cholesky_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_max_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_clamp_min_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_column_stack_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_conj_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_conj_physical_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_constant_pad_nd_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_copysign_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_copysign_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_corrcoef_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_corrcoef_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cos_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cosh_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cosh_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cosh_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cov_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cov_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cov_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummax_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummax_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cummin_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cumprod_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cumprod_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_cumprod_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_deg2rad_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_deg2rad_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diag_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diag_embed_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagflat_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagflat_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagonal_copy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagonal_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagonal_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagonal_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagonal_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_diagonal_scatter_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_digamma_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_double_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_dstack_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_permuted_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_empty_permuted_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_equal_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_equal_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_erf_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_erfc_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_erfinv_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_exp2_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_as_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expand_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expm1_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expm1_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_expm1_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_exponential_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fft2_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_fft2_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_hfft_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_hfft_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifft2_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ifft_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfftn_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfftn_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_ihfftn_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_irfft2_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_irfft_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfft2_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fft_rfft_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fill_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_flatten_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_flip_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_flip_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fliplr_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fliplr_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_float_power_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_floor_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fmax_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_fmod_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_full_like_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gather_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_geometric_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_geometric_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_gt_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_histc_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hsplit_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hstack_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_hstack_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_i0_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_add_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_fill_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_reduce_amax_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_reduce_prod_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_index_select_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_int_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_int_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isfinite_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isnan_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isnan_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_isneginf_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_2inputs_2outputs_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_2inputs_2outputs_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_2inputs_2outputs_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_2inputs_2outputs_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_2inputs_2outputs_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_4inputs_with_extra_args_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_4inputs_with_extra_args_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_jiterator_unary_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_kron_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lcm_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lcm_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ldexp_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ldexp_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lgamma_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lgamma_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_cholesky_ex_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_det_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_diagonal_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_diagonal_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_householder_product_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_inv_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_ldl_solve_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_ldl_solve_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_norm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_matrix_power_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_pinv_hermitian_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_pinv_singular_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_solve_ex_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linalg_solve_triangular_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linspace_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linspace_tensor_overload_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linspace_tensor_overload_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_linspace_tensor_overload_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log10_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log10_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log1p_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log1p_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log2_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log2_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_log_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logdet_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_and_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_and_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_and_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logical_or_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logspace_tensor_overload_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logspace_tensor_overload_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logsumexp_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_logsumexp_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lu_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lu_solve_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_lu_unpack_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mH_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mT_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_amin_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_amin_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_argmax_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_argmax_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_argmin_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_log_softmax_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_log_softmax_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_log_softmax_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_median_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_norm_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_normalize_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_normalize_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_scatter_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_select_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_select_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_masked_softmin_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_matrix_exp_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_maximum_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mean_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_meshgrid_list_of_tensors_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_meshgrid_variadic_tensors_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_min_reduction_with_dim_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_minimum_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mm_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mode_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_msort_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_msort_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_1_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_3_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_3_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_3_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_mvlgamma_mvlgamma_p_5_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nan_to_num_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nansum_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_narrow_copy_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_narrow_copy_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_narrow_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_narrow_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_native_batch_norm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ne_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_neg_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_neg_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_neg_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_empty_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_empty_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_empty_strided_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_full_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_full_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_ones_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_ones_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_new_zeros_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nextafter_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_adaptive_max_pool3d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_batch_norm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_batch_norm_without_cudnn_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_celu_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_channel_shuffle_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_channel_shuffle_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_conv_transpose2d_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_cosine_similarity_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_cross_entropy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_dropout2d_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_dropout2d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_embedding_bag_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_fractional_max_pool3d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_gelu_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_gelu_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_group_norm_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_group_norm_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_hardtanh_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_huber_loss_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_area_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_bicubic_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_l1_loss_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_linear_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_margin_ranking_loss_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_pool2d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_unpool1d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_max_unpool2d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_multi_head_attention_forward_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_multi_head_attention_forward_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_normalize_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_circular_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_reflect_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pad_replicate_negative_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_unshuffle_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_pixel_unshuffle_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_poisson_nll_loss_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_poisson_nll_loss_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_poisson_nll_loss_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_prelu_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_relu_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_scaled_dot_product_attention_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_silu_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_tanhshrink_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_tanhshrink_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_triplet_margin_loss_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_unfold_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nonzero_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nonzero_static_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ones_like_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_outer_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_pca_lowrank_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_copy_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_permute_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_1_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_2_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_2_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_3_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_3_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_polygamma_polygamma_n_3_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_pow_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_put_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_put_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rad2deg_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rad2deg_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_randint_like_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_ravel_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reciprocal_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reciprocal_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reciprocal_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reciprocal_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_remainder_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_repeat_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_repeat_interleave_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_repeat_interleave_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reshape_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_reshape_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resize__cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resize_as__cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resize_as__cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resize_as__cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_resolve_conj_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_roll_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rot90_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_round_decimals_0_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rsqrt_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_rsqrt_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_add_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_mean_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_scatter_reduce_sum_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_searchsorted_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_short_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sigmoid_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sign_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sign_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signal_windows_cosine_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_signal_windows_hamming_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinc_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sinh_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_slice_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_slice_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_slice_scatter_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_softmax_with_dtype_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sort_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sort_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_airy_ai_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_bessel_y0_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_chebyshev_polynomial_v_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_entr_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_entr_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_erfcx_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_hermite_polynomial_h_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_hermite_polynomial_h_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_i1_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_laguerre_polynomial_l_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_legendre_polynomial_p_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_log_ndtr_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_log_ndtr_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_modified_bessel_i1_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_modified_bessel_i1_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_modified_bessel_k0_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_modified_bessel_k0_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_modified_bessel_k1_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_polygamma_special_polygamma_n_0_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_scaled_modified_bessel_k1_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_u_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_v_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_xlog1py_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_comprehensive_special_xlog1py_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_split_with_sizes_copy_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_square_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_copy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_squeeze_multiple_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_std_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_std_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_std_mean_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_std_unbiased_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_stft_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_sum_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_svd_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_t_copy_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_t_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tan_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tan_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tanh_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_to_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_topk_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_topk_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_transpose_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapezoid_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapz_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trapz_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_triangular_solve_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tril_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_tril_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_triu_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_triu_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_true_divide_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_trunc_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unbind_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unbind_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unflatten_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unfold_copy_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unfold_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unfold_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unfold_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_uniform_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unique_consecutive_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unique_consecutive_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unique_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unique_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unravel_index_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsafe_split_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsafe_split_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_unsqueeze_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_mean_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_var_mean_unbiased_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vdot_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_view_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vstack_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_comprehensive_vstack_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_where_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_xlogy_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zero__cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zero__cuda_int64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_zeros_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick__native_batch_norm_legit_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick__unsafe_masked_index_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick__unsafe_masked_index_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick__unsafe_masked_index_put_accumulate_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_abs_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_acos_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_acosh_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_add_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_addcmul_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_addcmul_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_addr_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_addr_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_alias_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_all_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_any_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_any_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_any_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_as_strided_scatter_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_as_strided_scatter_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_as_strided_scatter_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_asinh_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_atanh_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_bernoulli_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_and_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_bitwise_left_shift_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_bucketize_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_ceil_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_ceil_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_clamp_min_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_clone_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_clone_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_addcmul_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_diagonal_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_fill_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_norm_inf_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_rad2deg_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_sgn_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_core_backward_std_mean_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_cos_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_cos_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_cosh_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_cosh_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_count_nonzero_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_cumprod_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_cumsum_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_diag_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_copy_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_diagonal_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_div_no_rounding_mode_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_empty_like_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_empty_like_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_empty_strided_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_empty_strided_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_eq_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_eq_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_erfc_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_erfinv_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_exp_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_expand_copy_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_expand_copy_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_expand_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_expand_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_expm1_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fft_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_fft_fft_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_hfft2_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_fft_hfftn_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ihfft2_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ihfft_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ihfft_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_fft_ihfftn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_irfftn_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_fft_rfft2_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_fft_rfftn_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_fill_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_flip_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_flip_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_floor_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_fmod_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_full_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_gcd_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_ge_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_geometric_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_grid_sampler_2d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_gt_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_gt_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_heaviside_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_index_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_index_fill_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_isin_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_isinf_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_isinf_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_isneginf_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_isneginf_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_item_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_cross_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_cross_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_cross_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_cross_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_cross_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_cross_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_linalg_diagonal_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_linspace_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_linspace_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_linspace_tensor_overload_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_linspace_tensor_overload_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_log10_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_log10_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_log10_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_logaddexp2_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_logical_not_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_logical_or_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_logical_xor_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_logical_xor_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_logit_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_logspace_tensor_overload_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_logspace_tensor_overload_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_logspace_tensor_overload_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_logsumexp_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_lt_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_masked_fill_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_mean_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_meshgrid_list_of_tensors_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_meshgrid_list_of_tensors_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_meshgrid_variadic_tensors_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_meshgrid_variadic_tensors_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_meshgrid_variadic_tensors_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_mul_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_mvlgamma_mvlgamma_p_1_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_mvlgamma_mvlgamma_p_1_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_mvlgamma_mvlgamma_p_3_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_mvlgamma_mvlgamma_p_5_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_nan_to_num_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_nansum_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_ne_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_ne_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_new_empty_strided_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_new_full_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_new_ones_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_new_ones_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_new_ones_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_elu_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_embedding_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_hardswish_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_hardtanh_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_max_unpool2d_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_max_unpool3d_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_max_unpool3d_grad_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_mish_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_relu_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_nn_functional_silu_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_norm_fro_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_normal_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_normal_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_normal_number_mean_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_ones_like_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_ones_like_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_permute_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_pow_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_pow_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_prod_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_prod_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_rad2deg_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_randn_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_reciprocal_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_remainder_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_renorm_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_renorm_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_repeat_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_repeat_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_roll_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_rot90_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_round_decimals_3_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_rsqrt_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_rsub_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_rsub_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_rsub_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_select_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_select_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_sigmoid_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_sign_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_sign_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_slice_scatter_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_special_entr_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_special_erfcx_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_special_log_ndtr_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_special_ndtri_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_split_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_split_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_split_list_args_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_split_with_sizes_copy_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_sqrt_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_squeeze_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_std_cuda_complex128, test/test_decomp.py::TestDecompCUDA::test_quick_std_mean_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_sub_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_sub_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_sub_cuda_int32, test/test_decomp.py::TestDecompCUDA::test_quick_sum_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_t_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_t_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_t_copy_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_t_copy_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_t_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_t_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_take_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_take_cuda_int64, test/test_decomp.py::TestDecompCUDA::test_quick_tanh_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_trace_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_trace_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_transpose_copy_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_transpose_copy_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_transpose_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_tril_cuda_int16, test/test_decomp.py::TestDecompCUDA::test_quick_triu_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_triu_cuda_float16, test/test_decomp.py::TestDecompCUDA::test_quick_unbind_copy_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_unbind_copy_cuda_complex32, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_unfold_cuda_int8, test/test_decomp.py::TestDecompCUDA::test_quick_uniform_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_unsafe_split_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_unsqueeze_copy_cuda_complex64, test/test_decomp.py::TestDecompCUDA::test_quick_unsqueeze_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_unsqueeze_cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_var_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_quick_var_mean_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_var_unbiased_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_view_copy_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_view_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_quick_view_cuda_uint8, test/test_decomp.py::TestDecompCUDA::test_quick_zero__cuda_bfloat16, test/test_decomp.py::TestDecompCUDA::test_quick_zero__cuda_bool, test/test_decomp.py::TestDecompCUDA::test_quick_zeros_like_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_rnn_decomp_module_nn_GRU_train_mode_cuda_float32, test/test_decomp.py::TestDecompCUDA::test_rnn_decomp_module_nn_RNN_eval_mode_cuda_float64, test/test_decomp.py::DecompOneOffTestsCUDA::test_contiguous_softmax_cuda 2025-07-17T09:35:56.0988219Z 2025-07-17T09:35:56.0988353Z Running test_mkldnn 1/1 ... [2025-07-17 09:35:56.073328] 2025-07-17T09:35:56.0988615Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:35:56.0989396Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_mkldnn.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:35:56.073621] 2025-07-17T09:35:59.9534109Z 2025-07-17T09:35:59.9535088Z test_mkldnn 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_mkldnn_1.1_24d0ac651d2a77fe_.log 2025-07-17T09:35:59.9535558Z Running 0 items in this shard: 2025-07-17T09:35:59.9535687Z 2025-07-17T09:35:59.9535851Z Running inductor/test_torchinductor 2/2 ... [2025-07-17 09:35:59.953222] 2025-07-17T09:35:59.9536152Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:35:59.9536862Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor.py', '-m', 'not serial', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:35:59.953506] 2025-07-17T09:41:24.9707230Z 2025-07-17T09:41:24.9708252Z test_jit_fuser_te 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_jit_fuser_te_1.1_ecc1c30a85deb151_.log 2025-07-17T09:41:25.1147798Z Running 6821 items in this shard: test/test_jit_fuser_te.py::TestFuserCommon::test_autodiff_fallback, test/test_jit_fuser_te.py::TestTEFuserStatic::test_abs, test/test_jit_fuser_te.py::TestTEFuserStatic::test_adaptive_avg_pool2d, test/test_jit_fuser_te.py::TestTEFuserStatic::test_add_bool, test/test_jit_fuser_te.py::TestTEFuserStatic::test_addcmul, test/test_jit_fuser_te.py::TestTEFuserStatic::test_arg_configurations_smoke, test/test_jit_fuser_te.py::TestTEFuserStatic::test_autocast_down, test/test_jit_fuser_te.py::TestTEFuserStatic::test_autocast_up, test/test_jit_fuser_te.py::TestTEFuserStatic::test_batch_norm, test/test_jit_fuser_te.py::TestTEFuserStatic::test_binary_div_ops, test/test_jit_fuser_te.py::TestTEFuserStatic::test_binary_ops, test/test_jit_fuser_te.py::TestTEFuserStatic::test_binary_pow, test/test_jit_fuser_te.py::TestTEFuserStatic::test_binary_scalar_ops, test/test_jit_fuser_te.py::TestTEFuserStatic::test_binary_tensor_scalar_ops, test/test_jit_fuser_te.py::TestTEFuserStatic::test_bitwise_ops, test/test_jit_fuser_te.py::TestTEFuserStatic::test_broadcast, test/test_jit_fuser_te.py::TestTEFuserStatic::test_cat_2k_args, test/test_jit_fuser_te.py::TestTEFuserStatic::test_cat_graph_opt, test/test_jit_fuser_te.py::TestTEFuserStatic::test_channels_last_dims_dynamic, test/test_jit_fuser_te.py::TestTEFuserStatic::test_checks_cat_inputs, test/test_jit_fuser_te.py::TestTEFuserStatic::test_chunk, test/test_jit_fuser_te.py::TestTEFuserStatic::test_chunk_correctness, test/test_jit_fuser_te.py::TestTEFuserStatic::test_chunk_distributes, test/test_jit_fuser_te.py::TestTEFuserStatic::test_chunk_motion_deduplicates_inputs, test/test_jit_fuser_te.py::TestTEFuserStatic::test_chunk_mul_one, test/test_jit_fuser_te.py::TestTEFuserStatic::test_chunk_multiple, test/test_jit_fuser_te.py::TestTEFuserStatic::test_clamp, test/test_jit_fuser_te.py::TestTEFuserStatic::test_clamp_double, test/test_jit_fuser_te.py::TestTEFuserStatic::test_clamp_int, test/test_jit_fuser_te.py::TestTEFuserStatic::test_comparison_eq_ne, test/test_jit_fuser_te.py::TestTEFuserStatic::test_comparison_ge_le, test/test_jit_fuser_te.py::TestTEFuserStatic::test_comparison_gt_lt, test/test_jit_fuser_te.py::TestTEFuserStatic::test_concat, test/test_jit_fuser_te.py::TestTEFuserStatic::test_concat_invariant, test/test_jit_fuser_te.py::TestTEFuserStatic::test_constant_chunk_shapes, test/test_jit_fuser_te.py::TestTEFuserStatic::test_conv2d, test/test_jit_fuser_te.py::TestTEFuserStatic::test_conv2d_depthwise, test/test_jit_fuser_te.py::TestTEFuserStatic::test_cuda_half, test/test_jit_fuser_te.py::TestTEFuserStatic::test_dims, test/test_jit_fuser_te.py::TestTEFuserStatic::test_disabled, test/test_jit_fuser_te.py::TestTEFuserStatic::test_div_bool, test/test_jit_fuser_te.py::TestTEFuserStatic::test_dynamic_cat, test/test_jit_fuser_te.py::TestTEFuserStatic::test_dynamic_shapes, test/test_jit_fuser_te.py::TestTEFuserStatic::test_eq_unsqueeze_type_as, test/test_jit_fuser_te.py::TestTEFuserStatic::test_erf, test/test_jit_fuser_te.py::TestTEFuserStatic::test_exhaust_specializations, test/test_jit_fuser_te.py::TestTEFuserStatic::test_exp, test/test_jit_fuser_te.py::TestTEFuserStatic::test_fusion_reuse_multi_gpu, test/test_jit_fuser_te.py::TestTEFuserStatic::test_gelu, test/test_jit_fuser_te.py::TestTEFuserStatic::test_hardsigmoid_fwd_bwd, test/test_jit_fuser_te.py::TestTEFuserStatic::test_hardswish_fwd_bwd, test/test_jit_fuser_te.py::TestTEFuserStatic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserStatic::test_isnan, test/test_jit_fuser_te.py::TestTEFuserStatic::test_kernel_cache_multi_gpu, test/test_jit_fuser_te.py::TestTEFuserStatic::test_lerp, test/test_jit_fuser_te.py::TestTEFuserStatic::test_list_ops, test/test_jit_fuser_te.py::TestTEFuserStatic::test_lstm, test/test_jit_fuser_te.py::TestTEFuserStatic::test_lstm_concat, test/test_jit_fuser_te.py::TestTEFuserStatic::test_lstm_gates_permutations, test/test_jit_fuser_te.py::TestTEFuserStatic::test_lstm_traced, test/test_jit_fuser_te.py::TestTEFuserStatic::test_masked_fill, test/test_jit_fuser_te.py::TestTEFuserStatic::test_matmul, test/test_jit_fuser_te.py::TestTEFuserStatic::test_milstm, test/test_jit_fuser_te.py::TestTEFuserStatic::test_minmax, test/test_jit_fuser_te.py::TestTEFuserStatic::test_minmax_int_ops, test/test_jit_fuser_te.py::TestTEFuserStatic::test_mul_bool, test/test_jit_fuser_te.py::TestTEFuserStatic::test_neg_pow, test/test_jit_fuser_te.py::TestTEFuserStatic::test_nonzero_device_cuda, test/test_jit_fuser_te.py::TestTEFuserStatic::test_nop, test/test_jit_fuser_te.py::TestTEFuserStatic::test_pow_multiple_dtype, test/test_jit_fuser_te.py::TestTEFuserStatic::test_profiler, test/test_jit_fuser_te.py::TestTEFuserStatic::test_rand_broadcast_cuda, test/test_jit_fuser_te.py::TestTEFuserStatic::test_rand_cuda, test/test_jit_fuser_te.py::TestTEFuserStatic::test_rand_diamond, test/test_jit_fuser_te.py::TestTEFuserStatic::test_relu, test/test_jit_fuser_te.py::TestTEFuserStatic::test_relu_fwd_bwd, test/test_jit_fuser_te.py::TestTEFuserStatic::test_remove_output_used_only_in_size, test/test_jit_fuser_te.py::TestTEFuserStatic::test_scalar, test/test_jit_fuser_te.py::TestTEFuserStatic::test_scalar_arg, test/test_jit_fuser_te.py::TestTEFuserStatic::test_scalar_only_inputs, test/test_jit_fuser_te.py::TestTEFuserStatic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserStatic::test_small_constant, test/test_jit_fuser_te.py::TestTEFuserStatic::test_sub_gt_and, test/test_jit_fuser_te.py::TestTEFuserStatic::test_sum_dim, test/test_jit_fuser_te.py::TestTEFuserStatic::test_sum_keepdim_cast, test/test_jit_fuser_te.py::TestTEFuserStatic::test_sum_simple, test/test_jit_fuser_te.py::TestTEFuserStatic::test_superslomo, test/test_jit_fuser_te.py::TestTEFuserStatic::test_tensor_scalar_ops, test/test_jit_fuser_te.py::TestTEFuserStatic::test_ternary_norm_ops, test/test_jit_fuser_te.py::TestTEFuserStatic::test_ternary_ops, test/test_jit_fuser_te.py::TestTEFuserStatic::test_threshold, test/test_jit_fuser_te.py::TestTEFuserStatic::test_to_device, test/test_jit_fuser_te.py::TestTEFuserStatic::test_to_dtype, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_type_as_cat, test/test_jit_fuser_te.py::TestTEFuserStatic::test_typecheck, test/test_jit_fuser_te.py::TestTEFuserStatic::test_unary_ops, test/test_jit_fuser_te.py::TestTEFuserStatic::test_unrolled_cat, test/test_jit_fuser_te.py::TestTEFuserStatic::test_unsqueeze_size_calculation, test/test_jit_fuser_te.py::TestTEFuserStatic::test_unsqueeze_var_dim, test/test_jit_fuser_te.py::TestTEFuserStatic::test_unsupported_dtypes, test/test_jit_fuser_te.py::TestTEFuserStatic::test_where_and_typing, test/test_jit_fuser_te.py::TestTEFuserStatic::test_where_ops, test/test_jit_fuser_te.py::TestTEFuserStatic::test_with_strict_fusion, test/test_jit_fuser_te.py::TestTEFuserStatic::test_zero_element_tensors, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_abs, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_adaptive_avg_pool2d, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_add_bool, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_addcmul, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_arg_configurations_smoke, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_autocast_down, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_autocast_up, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_batch_norm, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_binary_div_ops, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_binary_ops, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_binary_pow, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_binary_scalar_ops, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_binary_tensor_scalar_ops, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_bitwise_ops, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_broadcast, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_cat_2k_args, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_cat_graph_opt, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_channels_last_dims_dynamic, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_checks_cat_inputs, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_chunk, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_chunk_correctness, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_chunk_distributes, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_chunk_motion_deduplicates_inputs, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_chunk_mul_one, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_chunk_multiple, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_clamp, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_clamp_double, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_clamp_int, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_comparison_eq_ne, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_comparison_ge_le, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_comparison_gt_lt, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_concat, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_concat_invariant, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_constant_chunk_shapes, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_conv2d, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_conv2d_depthwise, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_cuda_half, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_dims, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_disabled, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_div_bool, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_dynamic_cat, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_dynamic_shapes, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_eq_unsqueeze_type_as, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_erf, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_exhaust_specializations, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_exp, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_fusion_reuse_multi_gpu, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_gelu, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_hardsigmoid_fwd_bwd, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_hardswish_fwd_bwd, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_isnan, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_kernel_cache_multi_gpu, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_lerp, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_list_ops, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_lstm, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_lstm_concat, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_lstm_gates_permutations, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_lstm_traced, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_masked_fill, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_matmul, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_milstm, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_minmax, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_minmax_int_ops, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_mul_bool, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_neg_pow, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_nonzero_device_cuda, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_nop, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_pow_multiple_dtype, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_profiler, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_rand_broadcast_cuda, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_rand_cuda, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_rand_diamond, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_relu, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_relu_fwd_bwd, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_remove_output_used_only_in_size, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_scalar, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_scalar_arg, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_scalar_only_inputs, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_small_constant, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_sub_gt_and, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_sum_dim, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_sum_keepdim_cast, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_sum_simple, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_superslomo, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_tensor_scalar_ops, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_ternary_norm_ops, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_ternary_ops, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_threshold, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_to_device, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_to_dtype, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_type_as_cat, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_typecheck, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_unary_ops, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_unrolled_cat, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_unsqueeze_size_calculation, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_unsqueeze_var_dim, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_unsupported_dtypes, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_where_and_typing, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_where_ops, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_with_strict_fusion, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_zero_element_tensors, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_failures___rmatmul___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_failures_frac_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_failures_matmul_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_H_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_T_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___getitem___cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___radd___cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___radd___cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___radd___cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___radd___cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___radd___cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___radd___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___radd___cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___radd___cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___radd___cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___radd___cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___radd___cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___radd___cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rand___cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rand___cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rand___cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rand___cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rand___cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rand___cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rdiv___cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rdiv___cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rdiv___cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rdiv___cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rdiv___cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rdiv___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rdiv___cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rdiv___cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rdiv___cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rdiv___cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rdiv___cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rdiv___cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmatmul___cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmatmul___cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmatmul___cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmatmul___cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmatmul___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmatmul___cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmod___cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmod___cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmod___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmod___cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmod___cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmod___cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmod___cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmod___cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmod___cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmul___cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmul___cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmul___cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmul___cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmul___cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmul___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmul___cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmul___cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmul___cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmul___cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmul___cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rmul___cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___ror___cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___ror___cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___ror___cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___ror___cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___ror___cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___ror___cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rpow___cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rpow___cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rpow___cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rpow___cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rpow___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rpow___cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rpow___cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rpow___cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rpow___cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rpow___cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rpow___cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rsub___cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rsub___cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rsub___cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rsub___cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rsub___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rsub___cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rsub___cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rsub___cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rsub___cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rsub___cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rsub___cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rxor___cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rxor___cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rxor___cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rxor___cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rxor___cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness___rxor___cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__batch_norm_with_update_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__batch_norm_with_update_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__batch_norm_with_update_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__batch_norm_with_update_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__chunk_cat_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__native_batch_norm_legit_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__native_batch_norm_legit_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__native_batch_norm_legit_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__native_batch_norm_legit_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__segment_reduce_lengths_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__segment_reduce_lengths_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__segment_reduce_lengths_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__segment_reduce_lengths_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__segment_reduce_offsets_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__segment_reduce_offsets_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__segment_reduce_offsets_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__segment_reduce_offsets_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__softmax_backward_data_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__softmax_backward_data_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__softmax_backward_data_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__softmax_backward_data_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_put_accumulate_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_put_accumulate_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_put_accumulate_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_put_accumulate_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_put_accumulate_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_put_accumulate_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_put_accumulate_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_put_accumulate_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_put_accumulate_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_put_accumulate_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_put_accumulate_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__unsafe_masked_index_put_accumulate_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__upsample_bilinear2d_aa_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__upsample_bilinear2d_aa_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__upsample_bilinear2d_aa_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness__upsample_bilinear2d_aa_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_abs_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acos_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_acosh_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_add_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addbmm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addbmm_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addbmm_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addbmm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addbmm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addbmm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcdiv_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcdiv_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcdiv_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcdiv_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcdiv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcdiv_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcmul_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcmul_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcmul_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcmul_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcmul_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcmul_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcmul_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcmul_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcmul_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcmul_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addcmul_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmm_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmm_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmm_decomposed_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmm_decomposed_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmm_decomposed_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmm_decomposed_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmm_decomposed_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmm_decomposed_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmv_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmv_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmv_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmv_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addmv_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addr_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addr_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addr_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addr_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addr_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addr_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addr_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addr_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addr_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addr_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_addr_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_alias_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_all_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_all_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_all_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_all_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_all_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_all_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_all_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_all_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_all_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_all_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_all_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_all_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_allclose_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_allclose_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_allclose_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_allclose_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_allclose_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_allclose_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amax_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amax_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amax_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amax_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amax_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amax_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amin_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amin_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amin_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amin_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amin_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_amin_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_aminmax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_aminmax_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_aminmax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_aminmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_aminmax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_aminmax_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_aminmax_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_aminmax_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_aminmax_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_aminmax_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_angle_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_angle_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_angle_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_angle_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_angle_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_angle_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_angle_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_angle_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_angle_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_angle_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_angle_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_any_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_any_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_any_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_any_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_any_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_any_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_any_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_any_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_any_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_any_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_any_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_any_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_arange_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_arange_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_arange_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_arange_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_arange_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_arange_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_arange_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_arange_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_arange_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmax_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmax_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmax_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmax_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmax_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmin_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmin_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmin_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmin_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argmin_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argsort_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argsort_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argsort_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argsort_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argsort_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argsort_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argsort_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argsort_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argsort_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argsort_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argwhere_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argwhere_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argwhere_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argwhere_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argwhere_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argwhere_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argwhere_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argwhere_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argwhere_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argwhere_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argwhere_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_argwhere_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_partial_views_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_as_strided_scatter_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asin_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_asinh_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan2_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan2_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan2_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan2_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan2_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan2_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan2_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan2_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan2_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atan_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atanh_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_1d_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_2d_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_atleast_3d_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_baddbmm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_baddbmm_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_baddbmm_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_baddbmm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_baddbmm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_baddbmm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bernoulli_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bernoulli_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bernoulli_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bernoulli_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bfloat16_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bincount_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bincount_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bincount_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bincount_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bincount_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_and_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_and_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_and_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_and_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_and_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_and_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_left_shift_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_left_shift_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_left_shift_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_left_shift_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_left_shift_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_not_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_not_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_not_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_not_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_not_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_not_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_or_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_or_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_or_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_or_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_or_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_or_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_right_shift_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_right_shift_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_right_shift_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_right_shift_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_right_shift_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_xor_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_xor_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_xor_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_xor_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_xor_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bitwise_xor_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_block_diag_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bmm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bmm_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bmm_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bmm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bmm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bmm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bool_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_shapes_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_tensors_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_tensors_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_tensors_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_tensors_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_tensors_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_tensors_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_tensors_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_tensors_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_tensors_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_tensors_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_tensors_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_tensors_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_to_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_to_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_to_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_to_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_to_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_to_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_to_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_to_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_to_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_to_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_to_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_broadcast_to_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bucketize_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bucketize_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bucketize_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bucketize_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bucketize_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bucketize_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bucketize_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bucketize_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_bucketize_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_byte_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_byte_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_byte_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_byte_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_byte_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_byte_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_byte_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_byte_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_byte_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_byte_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_byte_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_byte_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cartesian_prod_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cartesian_prod_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cartesian_prod_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cartesian_prod_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cartesian_prod_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cartesian_prod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cartesian_prod_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cartesian_prod_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cartesian_prod_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cartesian_prod_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cartesian_prod_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cartesian_prod_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cat_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cauchy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cauchy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cauchy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cauchy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdist_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdist_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cdouble_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ceil_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ceil_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ceil_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ceil_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ceil_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ceil_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ceil_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ceil_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ceil_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cfloat_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chalf_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_char_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cholesky_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cholesky_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cholesky_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cholesky_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cholesky_inverse_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cholesky_inverse_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cholesky_inverse_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cholesky_inverse_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cholesky_solve_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cholesky_solve_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cholesky_solve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cholesky_solve_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_chunk_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_max_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_max_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_max_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_max_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_max_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_max_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_max_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_max_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_max_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_max_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_min_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_min_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_min_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_min_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_min_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_min_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_min_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_min_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_min_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clamp_min_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_clone_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_column_stack_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_combinations_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_combinations_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_combinations_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_combinations_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_combinations_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_combinations_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_combinations_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_combinations_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_combinations_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_combinations_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_combinations_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_combinations_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_complex_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_complex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_complex_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_conj_physical_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_constant_pad_nd_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_constant_pad_nd_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_constant_pad_nd_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_constant_pad_nd_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_constant_pad_nd_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_constant_pad_nd_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_constant_pad_nd_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_constant_pad_nd_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_constant_pad_nd_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_constant_pad_nd_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_constant_pad_nd_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_constant_pad_nd_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_contiguous_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_copysign_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_copysign_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_copysign_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_copysign_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_copysign_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_copysign_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_copysign_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_copysign_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_copysign_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_copysign_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_corrcoef_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_corrcoef_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_corrcoef_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_corrcoef_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_corrcoef_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_corrcoef_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_corrcoef_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_corrcoef_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_corrcoef_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_corrcoef_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_corrcoef_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cos_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cosh_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_count_nonzero_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_count_nonzero_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_count_nonzero_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_count_nonzero_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_count_nonzero_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_count_nonzero_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_count_nonzero_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_count_nonzero_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_count_nonzero_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_count_nonzero_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_count_nonzero_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_count_nonzero_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cov_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cov_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cov_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cov_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cov_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cov_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cov_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cov_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cov_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cov_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cov_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cross_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cross_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cross_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cross_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cross_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cross_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cross_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cross_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cross_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cross_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cross_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummax_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummax_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummax_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummax_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummax_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummax_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummin_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummin_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummin_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummin_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummin_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cummin_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumprod_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumprod_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumprod_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumprod_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumprod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumprod_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumprod_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumprod_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumprod_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumprod_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumprod_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumsum_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumsum_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumsum_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumsum_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumsum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumsum_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumsum_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumsum_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumsum_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumsum_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumsum_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumulative_trapezoid_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumulative_trapezoid_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumulative_trapezoid_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumulative_trapezoid_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumulative_trapezoid_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumulative_trapezoid_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumulative_trapezoid_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumulative_trapezoid_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumulative_trapezoid_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumulative_trapezoid_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_cumulative_trapezoid_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_deg2rad_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_deg2rad_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_deg2rad_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_deg2rad_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_deg2rad_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_deg2rad_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_deg2rad_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_deg2rad_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_deg2rad_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_deg2rad_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diag_embed_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagflat_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagflat_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagflat_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagflat_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagflat_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagflat_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagflat_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagflat_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagflat_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagflat_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagflat_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagflat_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_scatter_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_scatter_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_scatter_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_scatter_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_scatter_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_scatter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_scatter_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_scatter_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_scatter_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_scatter_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_scatter_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diagonal_scatter_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diff_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diff_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diff_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diff_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diff_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diff_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diff_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diff_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diff_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diff_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diff_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_diff_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_digamma_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_digamma_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_digamma_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_digamma_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_digamma_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_digamma_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_digamma_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_digamma_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_digamma_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_digamma_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dist_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dist_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dist_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dist_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dist_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dist_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_floor_rounding_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_floor_rounding_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_floor_rounding_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_floor_rounding_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_floor_rounding_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_floor_rounding_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_floor_rounding_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_floor_rounding_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_floor_rounding_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_no_rounding_mode_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_trunc_rounding_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_trunc_rounding_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_trunc_rounding_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_trunc_rounding_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_trunc_rounding_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_trunc_rounding_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_trunc_rounding_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_trunc_rounding_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_div_trunc_rounding_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dot_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dot_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dot_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dot_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dot_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dot_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_double_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dsplit_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_dstack_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_einsum_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_einsum_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_einsum_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_einsum_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_einsum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_einsum_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_like_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_permuted_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_strided_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_strided_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_strided_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_strided_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_strided_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_strided_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_strided_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_strided_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_strided_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_strided_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_strided_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_empty_strided_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eq_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_equal_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_equal_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_equal_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_equal_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_equal_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_equal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_equal_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_equal_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_equal_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_equal_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_equal_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_equal_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erf_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erf_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erf_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erf_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erf_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erf_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erf_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erf_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erf_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfc_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfc_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfc_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfc_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfc_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfc_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfc_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfc_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfc_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfc_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfinv_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfinv_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfinv_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfinv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfinv_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfinv_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfinv_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfinv_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfinv_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_erfinv_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp2_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp2_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp2_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp2_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp2_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp2_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp2_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp2_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp2_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp2_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp2_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exp_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_as_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_as_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_as_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_as_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_as_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_as_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_as_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_as_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_as_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_as_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_as_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_as_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expand_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expm1_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expm1_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expm1_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expm1_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expm1_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expm1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expm1_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expm1_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expm1_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expm1_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expm1_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_expm1_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exponential_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exponential_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exponential_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_exponential_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_float8_e4m3fn, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_float8_e4m3fnuz, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_float8_e5m2, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_float8_e5m2fnuz, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_eye_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft2_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft2_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft2_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft2_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft2_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft2_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft2_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft2_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft2_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft2_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft2_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fft_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftn_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftn_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftn_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftn_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftn_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftn_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftn_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftn_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftn_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftn_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftn_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_fftshift_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft2_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft2_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft2_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft2_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft2_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft2_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft2_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft2_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft2_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft2_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft2_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfft_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfftn_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfftn_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfftn_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfftn_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfftn_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfftn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfftn_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfftn_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfftn_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfftn_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfftn_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_hfftn_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft2_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft2_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft2_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft2_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft2_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft2_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft2_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft2_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft2_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft2_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft2_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifft_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftn_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftn_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftn_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftn_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftn_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftn_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftn_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftn_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftn_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftn_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftn_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ifftshift_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft2_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft2_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft2_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft2_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft2_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft2_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft2_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft2_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfft_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfftn_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfftn_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfftn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfftn_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfftn_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfftn_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfftn_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfftn_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_ihfftn_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft2_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft2_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft2_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft2_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft2_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft2_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft2_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft2_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft2_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft2_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft2_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfft_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfftn_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfftn_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfftn_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfftn_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfftn_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfftn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfftn_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfftn_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfftn_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfftn_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfftn_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_irfftn_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft2_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft2_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft2_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft2_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft2_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft2_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft2_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft2_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfft_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfftn_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfftn_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfftn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfftn_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfftn_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfftn_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfftn_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfftn_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fft_rfftn_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fill_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flatten_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flip_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flip_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flip_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flip_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flip_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flip_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flip_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flip_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flip_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flip_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flip_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flip_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fliplr_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fliplr_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fliplr_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fliplr_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fliplr_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fliplr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fliplr_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fliplr_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fliplr_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fliplr_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fliplr_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fliplr_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flipud_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flipud_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flipud_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flipud_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flipud_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flipud_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flipud_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flipud_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flipud_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flipud_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flipud_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_flipud_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_power_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_power_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_power_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_power_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_power_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_power_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_power_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_power_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_power_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_power_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_power_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_float_power_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_divide_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_divide_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_divide_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_divide_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_divide_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_divide_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_divide_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_divide_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_floor_divide_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmax_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmax_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmax_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmax_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmax_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmax_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmin_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmin_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmin_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmin_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmin_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmin_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmod_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmod_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmod_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmod_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmod_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmod_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmod_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_fmod_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_frac_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_frac_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_frac_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_frac_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_frexp_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_frexp_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_frexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_frexp_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_uint16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_uint32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_full_like_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gather_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gather_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gather_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gather_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gather_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gather_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gather_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gather_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gather_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gather_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gather_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gather_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gcd_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gcd_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gcd_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gcd_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gcd_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ge_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ge_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ge_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ge_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ge_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ge_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ge_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ge_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ge_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ge_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geometric_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geometric_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geometric_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geometric_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geometric_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geometric_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geometric_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geometric_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geometric_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geqrf_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geqrf_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geqrf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_geqrf_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gradient_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gradient_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gradient_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gradient_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gradient_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gradient_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gradient_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gradient_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gradient_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gradient_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_grid_sampler_2d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_grid_sampler_2d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_grid_sampler_2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_grid_sampler_2d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gt_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gt_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gt_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gt_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gt_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gt_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gt_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gt_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gt_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_gt_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_half_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_half_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_half_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_half_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_half_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_half_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_half_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_half_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_half_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_half_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_half_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_half_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_heaviside_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_heaviside_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_heaviside_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_heaviside_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_heaviside_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_heaviside_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_heaviside_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_heaviside_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_heaviside_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_heaviside_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_histc_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_histc_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_histc_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_histc_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_histc_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_histc_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_histc_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hsplit_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hstack_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hypot_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hypot_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hypot_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_hypot_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_i0_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_i0_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_i0_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_i0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_i0_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_i0_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_i0_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_i0_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_i0_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_i0_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_igamma_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_igamma_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_igammac_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_igammac_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_imag_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_imag_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_imag_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_add_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_fill_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_put_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amax_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amax_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amax_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amax_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amax_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amin_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amin_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amin_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amin_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_amin_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_mean_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_mean_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_mean_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_mean_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_mean_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_mean_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_mean_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_mean_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_prod_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_prod_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_prod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_prod_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_prod_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_prod_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_prod_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_prod_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_reduce_prod_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_index_select_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_inner_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_inner_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_inner_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_inner_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_inner_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_inner_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_int_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_int_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_int_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_int_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_int_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_int_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_int_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_int_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_int_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_int_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_int_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_int_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isclose_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isclose_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isclose_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isclose_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isclose_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isclose_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isclose_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isclose_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isclose_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isclose_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isclose_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isclose_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isfinite_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isin_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isin_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isin_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isin_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isin_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isinf_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isnan_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isnan_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isnan_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isnan_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isnan_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isnan_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isnan_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isnan_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isnan_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isnan_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isnan_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isnan_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isneginf_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isneginf_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isneginf_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isneginf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isneginf_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isneginf_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isneginf_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isneginf_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isneginf_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isneginf_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isposinf_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isposinf_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isposinf_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isposinf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isposinf_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isposinf_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isposinf_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isposinf_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isposinf_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isposinf_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_isreal_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_istft_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_istft_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_item_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_2inputs_2outputs_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_2inputs_2outputs_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_2inputs_2outputs_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_2inputs_2outputs_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_2inputs_2outputs_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_2inputs_2outputs_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_2inputs_2outputs_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_2inputs_2outputs_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_2inputs_2outputs_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_2inputs_2outputs_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_2inputs_2outputs_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_2inputs_2outputs_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_4inputs_with_extra_args_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_4inputs_with_extra_args_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_4inputs_with_extra_args_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_4inputs_with_extra_args_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_4inputs_with_extra_args_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_4inputs_with_extra_args_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_4inputs_with_extra_args_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_4inputs_with_extra_args_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_4inputs_with_extra_args_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_4inputs_with_extra_args_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_4inputs_with_extra_args_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_4inputs_with_extra_args_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_return_by_ref_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_return_by_ref_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_return_by_ref_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_return_by_ref_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_return_by_ref_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_return_by_ref_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_return_by_ref_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_return_by_ref_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_return_by_ref_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_return_by_ref_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_return_by_ref_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_binary_return_by_ref_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_unary_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_unary_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_unary_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_unary_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_unary_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_unary_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_unary_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_unary_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_unary_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_unary_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_unary_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_jiterator_unary_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kron_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kron_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kron_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kron_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kron_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kron_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kron_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kron_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kron_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kron_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kron_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kron_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kthvalue_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kthvalue_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kthvalue_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kthvalue_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kthvalue_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kthvalue_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kthvalue_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kthvalue_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_kthvalue_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lcm_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lcm_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lcm_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lcm_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lcm_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ldexp_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ldexp_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ldexp_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ldexp_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ldexp_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ldexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ldexp_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ldexp_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ldexp_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ldexp_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ldexp_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ldexp_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_le_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_le_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_le_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_le_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_le_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_le_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_le_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_le_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_le_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_le_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lerp_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lerp_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lerp_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lerp_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lerp_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lerp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lerp_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lgamma_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lgamma_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lgamma_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lgamma_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lgamma_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lgamma_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lgamma_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lgamma_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lgamma_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lgamma_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cholesky_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cholesky_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cholesky_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cholesky_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cholesky_ex_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cholesky_ex_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cholesky_ex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cholesky_ex_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cond_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cond_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cond_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cond_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cross_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cross_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cross_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cross_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cross_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cross_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cross_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cross_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cross_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cross_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_cross_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_det_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_det_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_det_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_det_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_diagonal_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eig_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eig_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eig_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eig_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eigh_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eigh_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eigh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eigh_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eigvals_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eigvals_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eigvals_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eigvals_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eigvalsh_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eigvalsh_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eigvalsh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_eigvalsh_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_householder_product_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_householder_product_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_householder_product_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_householder_product_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_inv_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_inv_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_inv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_inv_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_inv_ex_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_inv_ex_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_inv_ex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_inv_ex_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_ldl_factor_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_ldl_factor_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_ldl_factor_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_ldl_factor_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_ldl_factor_ex_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_ldl_factor_ex_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_ldl_factor_ex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_ldl_factor_ex_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_ldl_solve_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_ldl_solve_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_ldl_solve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_ldl_solve_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lstsq_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lstsq_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lstsq_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lstsq_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lstsq_grad_oriented_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lstsq_grad_oriented_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lstsq_grad_oriented_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lstsq_grad_oriented_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_factor_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_factor_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_factor_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_factor_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_factor_ex_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_factor_ex_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_factor_ex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_factor_ex_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_solve_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_solve_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_solve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_lu_solve_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_norm_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_norm_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_power_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_power_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_power_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_power_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_rank_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_rank_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_rank_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_rank_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_rank_hermitian_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_rank_hermitian_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_rank_hermitian_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_matrix_rank_hermitian_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_multi_dot_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_multi_dot_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_multi_dot_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_multi_dot_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_multi_dot_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_multi_dot_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_norm_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_norm_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_norm_subgradients_at_zero_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_norm_subgradients_at_zero_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_norm_subgradients_at_zero_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_norm_subgradients_at_zero_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_norm_subgradients_at_zero_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_norm_subgradients_at_zero_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_pinv_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_pinv_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_pinv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_pinv_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_pinv_hermitian_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_pinv_hermitian_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_pinv_hermitian_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_pinv_hermitian_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_pinv_singular_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_pinv_singular_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_pinv_singular_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_pinv_singular_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_qr_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_qr_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_qr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_qr_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_slogdet_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_slogdet_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_slogdet_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_slogdet_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_solve_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_solve_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_solve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_solve_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_solve_ex_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_solve_ex_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_solve_ex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_solve_ex_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_solve_triangular_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_solve_triangular_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_solve_triangular_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_solve_triangular_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_svd_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_svd_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_svd_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_svd_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_svdvals_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_svdvals_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_svdvals_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_svdvals_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_tensorinv_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_tensorinv_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_tensorinv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_tensorinv_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_tensorsolve_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_tensorsolve_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_tensorsolve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_tensorsolve_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vander_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vander_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vander_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vander_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vander_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vander_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vander_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vander_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vander_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vecdot_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vecdot_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vecdot_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vecdot_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vecdot_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vecdot_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vector_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vector_norm_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vector_norm_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vector_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vector_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linalg_vector_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_tensor_overload_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_tensor_overload_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_tensor_overload_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_tensor_overload_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_tensor_overload_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_tensor_overload_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_tensor_overload_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_tensor_overload_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_tensor_overload_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_tensor_overload_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_linspace_tensor_overload_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log10_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log10_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log10_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log10_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log10_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log10_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log10_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log10_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log10_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log10_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log10_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log10_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log1p_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log1p_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log1p_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log1p_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log1p_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log1p_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log1p_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log1p_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log1p_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log1p_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log1p_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log1p_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log2_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log2_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log2_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log2_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log2_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log2_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log2_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log2_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log2_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log2_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log2_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_normal_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_normal_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_normal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_normal_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_log_softmax_with_dtype_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logaddexp2_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logaddexp2_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logaddexp2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logaddexp2_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logaddexp_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logaddexp_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logaddexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logaddexp_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logcumsumexp_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logcumsumexp_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logcumsumexp_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logcumsumexp_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logcumsumexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logcumsumexp_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logdet_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logdet_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logdet_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logdet_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_and_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_and_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_and_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_and_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_and_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_and_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_and_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_and_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_and_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_and_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_and_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_and_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_not_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_not_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_not_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_not_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_not_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_not_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_not_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_not_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_not_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_not_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_not_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_not_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_or_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_or_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_or_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_or_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_or_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_or_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_or_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_or_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_or_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_or_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_or_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_or_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_xor_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_xor_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_xor_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_xor_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_xor_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_xor_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_xor_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_xor_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_xor_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_xor_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_xor_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logical_xor_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logit_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logit_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logit_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logit_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logit_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logit_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logit_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logit_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logit_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logit_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_tensor_overload_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_tensor_overload_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_tensor_overload_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_tensor_overload_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_tensor_overload_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_tensor_overload_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_tensor_overload_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_tensor_overload_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_tensor_overload_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_tensor_overload_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logspace_tensor_overload_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logsumexp_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logsumexp_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logsumexp_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logsumexp_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logsumexp_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logsumexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logsumexp_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logsumexp_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logsumexp_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logsumexp_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logsumexp_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_logsumexp_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_long_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lt_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lt_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lt_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lt_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lt_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lt_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lt_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lt_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lt_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lt_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lu_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lu_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lu_solve_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lu_solve_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lu_solve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lu_solve_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lu_unpack_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lu_unpack_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lu_unpack_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_lu_unpack_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mH_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mT_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amax_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amax_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amax_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amax_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amax_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amin_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amin_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amin_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amin_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_amin_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmax_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmax_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmax_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmax_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmax_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmin_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmin_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmin_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmin_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_argmin_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumprod_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumprod_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumprod_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumprod_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumprod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumprod_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumprod_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumprod_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumprod_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumprod_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumprod_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumsum_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumsum_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumsum_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumsum_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumsum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumsum_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumsum_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumsum_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumsum_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumsum_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_cumsum_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_fill_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_log_softmax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_log_softmax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_log_softmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_log_softmax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logaddexp_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logaddexp_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logaddexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logaddexp_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logsumexp_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logsumexp_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logsumexp_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logsumexp_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logsumexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logsumexp_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logsumexp_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logsumexp_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logsumexp_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logsumexp_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_logsumexp_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_mean_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_mean_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_mean_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_mean_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_mean_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_median_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_median_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_median_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_median_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_normalize_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_normalize_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_normalize_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_normalize_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_normalize_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_normalize_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_prod_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_prod_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_prod_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_prod_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_prod_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_prod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_prod_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_prod_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_prod_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_prod_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_prod_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_prod_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_scatter_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_scatter_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_scatter_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_scatter_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_scatter_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_scatter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_scatter_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_scatter_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_scatter_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_scatter_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_scatter_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_scatter_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_select_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_select_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_select_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_select_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_select_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_select_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_select_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_select_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_select_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_select_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_select_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_select_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_softmax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_softmax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_softmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_softmax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_softmin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_softmin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_softmin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_softmin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_std_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_std_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_std_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_std_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_std_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_std_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_std_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_std_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_std_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_std_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_std_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_sum_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_sum_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_sum_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_sum_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_sum_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_sum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_sum_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_sum_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_sum_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_sum_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_sum_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_sum_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_var_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_var_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_var_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_var_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_var_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_var_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_var_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_var_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_var_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_var_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_masked_var_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_matmul_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_matmul_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_matmul_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_matmul_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_matmul_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_matmul_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_matrix_exp_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_matrix_exp_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_matrix_exp_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_matrix_exp_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_matrix_exp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_matrix_exp_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_binary_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_binary_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_binary_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_binary_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_binary_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_binary_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_binary_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_binary_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_binary_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_binary_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_pool2d_with_indices_backward_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_pool2d_with_indices_backward_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_pool2d_with_indices_backward_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_pool2d_with_indices_backward_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_no_dim_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_no_dim_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_no_dim_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_no_dim_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_no_dim_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_no_dim_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_no_dim_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_no_dim_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_no_dim_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_no_dim_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_with_dim_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_with_dim_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_with_dim_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_with_dim_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_with_dim_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_with_dim_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_with_dim_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_with_dim_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_with_dim_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_max_reduction_with_dim_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_maximum_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_maximum_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_maximum_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_maximum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_maximum_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_maximum_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_maximum_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_maximum_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_maximum_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_maximum_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mean_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mean_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mean_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mean_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mean_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_median_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_median_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_median_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_median_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_median_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_median_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_median_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_median_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_median_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_list_of_tensors_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_list_of_tensors_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_list_of_tensors_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_list_of_tensors_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_list_of_tensors_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_list_of_tensors_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_list_of_tensors_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_list_of_tensors_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_list_of_tensors_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_list_of_tensors_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_list_of_tensors_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_list_of_tensors_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_variadic_tensors_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_variadic_tensors_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_variadic_tensors_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_variadic_tensors_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_variadic_tensors_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_variadic_tensors_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_variadic_tensors_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_variadic_tensors_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_variadic_tensors_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_variadic_tensors_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_variadic_tensors_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_meshgrid_variadic_tensors_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_binary_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_binary_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_binary_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_binary_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_binary_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_binary_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_binary_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_binary_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_binary_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_binary_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_no_dim_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_no_dim_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_no_dim_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_no_dim_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_no_dim_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_no_dim_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_no_dim_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_no_dim_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_no_dim_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_no_dim_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_with_dim_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_with_dim_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_with_dim_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_with_dim_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_with_dim_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_with_dim_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_with_dim_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_with_dim_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_with_dim_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_min_reduction_with_dim_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_minimum_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_minimum_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_minimum_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_minimum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_minimum_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_minimum_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_minimum_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_minimum_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_minimum_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_minimum_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mm_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mm_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mode_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mode_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mode_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mode_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mode_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mode_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mode_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mode_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mode_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mode_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_movedim_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_msort_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_msort_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_msort_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_msort_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_msort_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_msort_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_msort_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_msort_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_msort_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_msort_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mul_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_multinomial_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_multinomial_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_multinomial_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_multinomial_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mv_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mv_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mv_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mv_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mv_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_1_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_1_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_1_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_1_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_1_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_1_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_1_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_1_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_3_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_3_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_3_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_3_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_3_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_3_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_3_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_3_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_3_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_5_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_5_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_5_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_5_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_5_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_5_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_5_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_mvlgamma_mvlgamma_p_5_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nan_to_num_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nan_to_num_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nan_to_num_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nan_to_num_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nan_to_num_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nan_to_num_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nan_to_num_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nan_to_num_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nan_to_num_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nan_to_num_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmean_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmean_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmean_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmean_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmean_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmean_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmedian_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmedian_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmedian_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmedian_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmedian_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmedian_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmedian_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmedian_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanmedian_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanquantile_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nanquantile_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nansum_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_narrow_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_native_batch_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_native_batch_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_native_batch_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_native_batch_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_native_dropout_backward_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_native_dropout_backward_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_native_dropout_backward_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_native_dropout_backward_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_native_layer_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_native_layer_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_native_layer_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_native_layer_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ne_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ne_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ne_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ne_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ne_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ne_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ne_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ne_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ne_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ne_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ne_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ne_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_neg_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_neg_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_neg_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_neg_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_neg_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_neg_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_neg_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_neg_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_neg_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_neg_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_neg_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_neg_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_empty_strided_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_full_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_ones_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_new_zeros_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nextafter_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nextafter_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nextafter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nextafter_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_avg_pool1d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_avg_pool1d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_avg_pool1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_avg_pool1d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_avg_pool2d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_avg_pool2d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_avg_pool2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_avg_pool2d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_avg_pool3d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_avg_pool3d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_avg_pool3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_avg_pool3d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_max_pool1d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_max_pool1d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_max_pool1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_max_pool1d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_max_pool2d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_max_pool2d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_max_pool2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_max_pool2d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_max_pool3d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_max_pool3d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_max_pool3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_adaptive_max_pool3d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_alpha_dropout_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_alpha_dropout_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_alpha_dropout_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_alpha_dropout_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_avg_pool1d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_avg_pool1d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_avg_pool1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_avg_pool1d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_avg_pool2d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_avg_pool2d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_avg_pool2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_avg_pool2d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_avg_pool3d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_avg_pool3d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_avg_pool3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_avg_pool3d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_batch_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_batch_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_batch_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_batch_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_batch_norm_without_cudnn_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_batch_norm_without_cudnn_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_batch_norm_without_cudnn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_batch_norm_without_cudnn_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_bilinear_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_bilinear_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_bilinear_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_bilinear_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_binary_cross_entropy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_binary_cross_entropy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_binary_cross_entropy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_binary_cross_entropy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_binary_cross_entropy_with_logits_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_binary_cross_entropy_with_logits_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_binary_cross_entropy_with_logits_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_celu_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_celu_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_celu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_celu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_channel_shuffle_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_channel_shuffle_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_channel_shuffle_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_channel_shuffle_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_channel_shuffle_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_channel_shuffle_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_channel_shuffle_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_channel_shuffle_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_channel_shuffle_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_channel_shuffle_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_channel_shuffle_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_channel_shuffle_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv1d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv1d_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv1d_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv1d_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv1d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv1d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv2d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv2d_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv2d_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv2d_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv2d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv2d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv3d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv3d_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv3d_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv3d_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv3d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv3d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose1d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose1d_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose1d_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose1d_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose1d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose1d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose2d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose2d_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose2d_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose2d_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose2d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose2d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose3d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose3d_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose3d_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose3d_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose3d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_conv_transpose3d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_embedding_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_embedding_loss_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_embedding_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_embedding_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_embedding_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_embedding_loss_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_embedding_loss_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_embedding_loss_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_embedding_loss_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_embedding_loss_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_similarity_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_similarity_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_similarity_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cosine_similarity_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cross_entropy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cross_entropy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cross_entropy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_cross_entropy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_ctc_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_ctc_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_dropout2d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_dropout2d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_dropout2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_dropout2d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_dropout3d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_dropout3d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_dropout3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_dropout3d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_dropout_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_dropout_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_dropout_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_dropout_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_elu_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_elu_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_elu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_elu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_embedding_bag_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_embedding_bag_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_embedding_bag_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_embedding_bag_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_embedding_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_embedding_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_embedding_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_embedding_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_with_train_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_with_train_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_with_train_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_without_train_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_without_train_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_without_train_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_without_train_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_without_train_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_without_train_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_without_train_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_without_train_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_without_train_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_without_train_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_feature_alpha_dropout_without_train_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_fractional_max_pool2d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_fractional_max_pool2d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_fractional_max_pool2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_fractional_max_pool2d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_fractional_max_pool3d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_fractional_max_pool3d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_fractional_max_pool3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_fractional_max_pool3d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_gaussian_nll_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_gaussian_nll_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_gaussian_nll_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_gaussian_nll_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_gelu_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_gelu_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_gelu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_gelu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_glu_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_glu_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_glu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_glu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_grid_sample_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_grid_sample_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_grid_sample_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_grid_sample_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_group_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_group_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_group_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_group_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardshrink_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardshrink_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardshrink_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardshrink_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardsigmoid_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardsigmoid_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardsigmoid_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardsigmoid_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardswish_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardswish_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardswish_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardswish_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardtanh_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardtanh_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardtanh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardtanh_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardtanh_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardtanh_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardtanh_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hardtanh_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hinge_embedding_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hinge_embedding_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hinge_embedding_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_hinge_embedding_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_huber_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_huber_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_huber_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_huber_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_instance_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_instance_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_instance_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_instance_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_area_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_area_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_area_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_area_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_bicubic_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_bicubic_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_bicubic_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_bicubic_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_bilinear_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_bilinear_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_bilinear_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_bilinear_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_linear_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_linear_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_linear_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_linear_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_nearest-exact_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_nearest-exact_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_nearest-exact_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_nearest-exact_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_nearest_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_nearest_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_nearest_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_nearest_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_nearest_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_trilinear_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_trilinear_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_trilinear_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_interpolate_trilinear_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_kl_div_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_kl_div_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_kl_div_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_kl_div_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_l1_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_l1_loss_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_l1_loss_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_l1_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_l1_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_l1_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_layer_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_layer_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_layer_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_layer_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_leaky_relu_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_leaky_relu_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_leaky_relu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_leaky_relu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_linear_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_linear_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_linear_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_linear_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_linear_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_linear_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_local_response_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_local_response_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_local_response_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_local_response_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_logsigmoid_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_logsigmoid_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_logsigmoid_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_logsigmoid_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_margin_ranking_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_margin_ranking_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_margin_ranking_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_margin_ranking_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_margin_ranking_loss_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_margin_ranking_loss_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_margin_ranking_loss_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_margin_ranking_loss_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_margin_ranking_loss_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_pool1d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_pool1d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_pool1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_pool1d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_pool2d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_pool2d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_pool2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_pool2d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_pool3d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_pool3d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_pool3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_pool3d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool1d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool1d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool1d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool1d_grad_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool1d_grad_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool1d_grad_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool1d_grad_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool2d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool2d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool2d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool2d_grad_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool2d_grad_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool2d_grad_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool2d_grad_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool3d_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool3d_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool3d_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool3d_grad_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool3d_grad_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool3d_grad_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_max_unpool3d_grad_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_mish_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_mish_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_mish_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_mish_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_mse_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_mse_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_mse_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_mse_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multi_head_attention_forward_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multi_head_attention_forward_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multi_head_attention_forward_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multi_head_attention_forward_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multi_margin_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multi_margin_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multi_margin_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multi_margin_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multilabel_margin_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multilabel_margin_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multilabel_margin_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multilabel_margin_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multilabel_soft_margin_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multilabel_soft_margin_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_multilabel_soft_margin_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_nll_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_nll_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_nll_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_nll_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_normalize_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_normalize_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_normalize_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_normalize_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_normalize_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_normalize_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_one_hot_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_circular_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_circular_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_circular_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_circular_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_circular_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_circular_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_circular_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_circular_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_circular_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_circular_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_circular_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_circular_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_constant_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_constant_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_constant_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_constant_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_constant_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_constant_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_constant_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_constant_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_constant_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_constant_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_constant_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_constant_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_reflect_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_reflect_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_reflect_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_reflect_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_reflect_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_reflect_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_reflect_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_reflect_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_reflect_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_reflect_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_reflect_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_negative_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_negative_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_negative_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_negative_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_negative_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_negative_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_negative_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_negative_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_negative_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_negative_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pad_replicate_negative_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pairwise_distance_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pairwise_distance_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pairwise_distance_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pairwise_distance_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pairwise_distance_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pairwise_distance_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pairwise_distance_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pairwise_distance_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pairwise_distance_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pairwise_distance_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pairwise_distance_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pdist_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pdist_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_shuffle_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_shuffle_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_shuffle_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_shuffle_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_shuffle_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_shuffle_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_shuffle_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_shuffle_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_shuffle_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_shuffle_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_shuffle_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_shuffle_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_unshuffle_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_unshuffle_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_unshuffle_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_unshuffle_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_unshuffle_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_unshuffle_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_unshuffle_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_unshuffle_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_unshuffle_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_unshuffle_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_unshuffle_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_pixel_unshuffle_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_poisson_nll_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_poisson_nll_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_poisson_nll_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_poisson_nll_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_poisson_nll_loss_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_poisson_nll_loss_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_poisson_nll_loss_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_poisson_nll_loss_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_poisson_nll_loss_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_prelu_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_prelu_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_prelu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_prelu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu6_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu6_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu6_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu6_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu6_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu6_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu6_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu6_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu6_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_relu_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_rms_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_rms_norm_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_rms_norm_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_rms_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_rms_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_rms_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_rrelu_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_rrelu_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_rrelu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_rrelu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_scaled_dot_product_attention_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_scaled_dot_product_attention_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_scaled_dot_product_attention_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_scaled_dot_product_attention_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_selu_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_selu_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_selu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_selu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_silu_complex_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_silu_complex_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_silu_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_silu_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_silu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_silu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_smooth_l1_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_smooth_l1_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_smooth_l1_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_smooth_l1_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_soft_margin_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_soft_margin_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_soft_margin_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_soft_margin_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_with_dtype_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_with_dtype_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_with_dtype_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_with_dtype_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_with_dtype_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_with_dtype_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_with_dtype_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_with_dtype_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_with_dtype_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_with_dtype_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softmin_with_dtype_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softplus_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softplus_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softplus_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softplus_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softshrink_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softshrink_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softshrink_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softshrink_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softsign_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softsign_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softsign_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softsign_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softsign_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softsign_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softsign_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softsign_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softsign_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softsign_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softsign_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_softsign_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_tanhshrink_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_tanhshrink_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_tanhshrink_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_tanhshrink_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_tanhshrink_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_tanhshrink_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_tanhshrink_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_tanhshrink_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_tanhshrink_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_tanhshrink_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_tanhshrink_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_threshold_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_threshold_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_threshold_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_threshold_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_threshold_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_threshold_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_threshold_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_threshold_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_threshold_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_loss_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_loss_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_loss_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_loss_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_loss_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_loss_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_loss_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_with_distance_loss_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_with_distance_loss_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_with_distance_loss_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_with_distance_loss_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_with_distance_loss_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_with_distance_loss_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_with_distance_loss_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_with_distance_loss_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_with_distance_loss_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_triplet_margin_with_distance_loss_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_unfold_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_unfold_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_unfold_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_unfold_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_unfold_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_unfold_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_unfold_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_upsample_bilinear_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_upsample_bilinear_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_upsample_bilinear_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_upsample_bilinear_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_upsample_nearest_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_upsample_nearest_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_upsample_nearest_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_upsample_nearest_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nn_functional_upsample_nearest_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_nonzero_static_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_fro_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_fro_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_fro_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_fro_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_fro_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_fro_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_inf_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_inf_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_inf_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_inf_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_inf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_inf_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_nuc_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_nuc_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_nuc_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_norm_nuc_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_in_place_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_in_place_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_in_place_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_in_place_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_in_place_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_in_place_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_number_mean_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_number_mean_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_number_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_normal_number_mean_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ones_like_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ormqr_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ormqr_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ormqr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ormqr_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_outer_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_outer_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_outer_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_outer_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_outer_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_outer_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_outer_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_outer_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_outer_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_outer_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_outer_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_outer_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pca_lowrank_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pca_lowrank_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pca_lowrank_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pca_lowrank_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_permute_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pinverse_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pinverse_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pinverse_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pinverse_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polar_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polar_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_0_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_0_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_0_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_0_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_0_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_0_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_0_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_0_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_0_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_1_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_1_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_1_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_1_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_1_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_1_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_1_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_1_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_1_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_2_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_2_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_2_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_2_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_2_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_2_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_2_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_2_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_2_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_3_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_3_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_3_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_3_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_3_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_3_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_3_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_3_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_3_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_3_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_4_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_4_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_4_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_4_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_4_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_4_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_4_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_4_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_4_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_polygamma_polygamma_n_4_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_positive_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_positive_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_positive_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_positive_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_positive_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_positive_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_positive_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_positive_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_positive_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_positive_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_positive_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_positive_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pow_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pow_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pow_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pow_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pow_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pow_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pow_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pow_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pow_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pow_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pow_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_pow_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_prod_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_put_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_put_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_put_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_put_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_put_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_put_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_put_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_put_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_put_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_put_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_put_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_put_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_qr_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_qr_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_qr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_qr_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_quantile_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_quantile_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rad2deg_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rad2deg_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rad2deg_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rad2deg_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rad2deg_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rad2deg_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rad2deg_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rad2deg_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rad2deg_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rad2deg_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rand_like_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rand_like_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rand_like_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rand_like_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rand_like_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rand_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rand_like_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_like_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_like_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_like_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_like_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_like_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_like_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_like_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randint_like_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_like_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_like_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_like_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_like_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_like_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_randn_like_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_ravel_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_real_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reciprocal_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reciprocal_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reciprocal_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reciprocal_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reciprocal_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reciprocal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reciprocal_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reciprocal_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reciprocal_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reciprocal_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reciprocal_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reciprocal_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_remainder_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_remainder_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_remainder_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_remainder_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_remainder_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_remainder_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_remainder_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_remainder_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_remainder_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_renorm_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_renorm_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_renorm_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_renorm_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_renorm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_renorm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_repeat_interleave_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_as_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_reshape_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize__cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize__cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize__cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize__cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize__cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize__cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize__cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize__cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize__cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize__cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize__cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize__cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize_as__cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize_as__cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize_as__cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize_as__cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize_as__cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize_as__cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize_as__cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize_as__cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize_as__cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize_as__cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize_as__cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resize_as__cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_conj_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_conj_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_conj_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_conj_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_conj_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_conj_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_conj_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_conj_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_conj_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_conj_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_conj_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_conj_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_resolve_neg_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_roll_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rot90_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rot90_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rot90_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rot90_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rot90_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rot90_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rot90_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rot90_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rot90_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rot90_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rot90_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rot90_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_decimals_0_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_decimals_0_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_decimals_0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_decimals_0_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_decimals_3_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_decimals_3_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_decimals_3_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_decimals_3_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_decimals_neg_3_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_decimals_neg_3_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_decimals_neg_3_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_round_decimals_neg_3_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsqrt_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsub_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsub_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsub_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsub_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsub_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsub_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsub_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsub_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsub_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsub_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_rsub_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scalar_tensor_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_add_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_add_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_add_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_add_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_add_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_add_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_add_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_add_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_add_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_add_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_add_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_add_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amax_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amax_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amax_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amax_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amax_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amin_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amin_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amin_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amin_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_amin_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_mean_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_mean_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_mean_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_mean_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_mean_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_mean_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_mean_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_mean_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_prod_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_prod_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_prod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_prod_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_prod_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_prod_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_prod_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_prod_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_prod_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_sum_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_sum_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_sum_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_sum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_sum_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_sum_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_sum_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_sum_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_sum_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_scatter_reduce_sum_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_searchsorted_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_searchsorted_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_searchsorted_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_searchsorted_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_searchsorted_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_searchsorted_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_searchsorted_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_searchsorted_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_searchsorted_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_scatter_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_scatter_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_scatter_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_scatter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_scatter_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_scatter_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_scatter_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_scatter_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_scatter_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_select_scatter_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sgn_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_short_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_short_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_short_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_short_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_short_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_short_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_short_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_short_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_short_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_short_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_short_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_short_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sigmoid_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sign_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sign_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sign_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sign_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sign_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sign_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sign_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sign_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sign_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sign_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_bartlett_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_bartlett_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_blackman_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_blackman_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_cosine_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_cosine_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_exponential_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_exponential_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_gaussian_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_gaussian_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_general_cosine_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_general_cosine_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_general_hamming_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_general_hamming_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_hamming_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_hamming_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_hann_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_hann_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_kaiser_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_kaiser_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_nuttall_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signal_windows_nuttall_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signbit_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signbit_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signbit_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signbit_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signbit_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signbit_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signbit_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signbit_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signbit_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_signbit_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sin_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinc_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinc_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinc_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinc_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinc_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinc_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinc_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinc_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinc_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinc_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinc_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinc_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sinh_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_scatter_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_scatter_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_scatter_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_scatter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_scatter_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_scatter_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_scatter_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_scatter_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_scatter_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_slice_scatter_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_with_dtype_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_with_dtype_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_with_dtype_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_with_dtype_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_with_dtype_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_with_dtype_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_with_dtype_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_with_dtype_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_with_dtype_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_with_dtype_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_with_dtype_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_softmax_with_dtype_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sort_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sort_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sort_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sort_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sort_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sort_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sort_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sort_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sort_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sort_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sparse_mm_reduce_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sparse_mm_reduce_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sparse_mm_reduce_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sparse_mm_reduce_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sparse_sampled_addmm_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sparse_sampled_addmm_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sparse_sampled_addmm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sparse_sampled_addmm_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_airy_ai_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_airy_ai_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_airy_ai_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_airy_ai_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_airy_ai_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_airy_ai_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_airy_ai_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_airy_ai_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j0_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j0_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j0_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j0_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j0_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j0_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j0_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j1_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j1_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j1_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j1_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j1_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j1_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_j1_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y0_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y0_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y0_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y0_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y0_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y0_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y0_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y1_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y1_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y1_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y1_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y1_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y1_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_bessel_y1_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_t_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_t_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_t_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_t_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_t_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_t_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_t_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_t_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_u_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_u_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_u_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_u_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_u_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_u_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_u_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_u_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_v_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_v_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_v_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_v_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_v_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_v_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_v_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_v_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_w_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_w_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_w_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_w_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_w_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_w_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_w_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_chebyshev_polynomial_w_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_entr_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_entr_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_entr_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_entr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_entr_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_entr_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_entr_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_entr_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_entr_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_entr_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_erfcx_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_erfcx_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_erfcx_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_erfcx_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_erfcx_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_erfcx_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_erfcx_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_erfcx_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_h_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_h_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_h_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_h_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_h_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_h_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_h_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_h_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_he_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_he_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_he_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_he_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_he_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_he_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_he_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_hermite_polynomial_he_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i0e_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i0e_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i0e_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i0e_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i0e_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i0e_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i0e_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i0e_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i0e_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i0e_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1e_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1e_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1e_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1e_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1e_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1e_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1e_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1e_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1e_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_i1e_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_laguerre_polynomial_l_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_laguerre_polynomial_l_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_laguerre_polynomial_l_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_laguerre_polynomial_l_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_laguerre_polynomial_l_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_laguerre_polynomial_l_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_laguerre_polynomial_l_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_laguerre_polynomial_l_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_legendre_polynomial_p_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_legendre_polynomial_p_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_legendre_polynomial_p_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_legendre_polynomial_p_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_legendre_polynomial_p_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_legendre_polynomial_p_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_legendre_polynomial_p_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_legendre_polynomial_p_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_log_ndtr_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_log_ndtr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_log_ndtr_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_log_ndtr_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_log_ndtr_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_log_ndtr_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_log_ndtr_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_log_ndtr_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i0_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i0_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i0_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i0_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i0_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i0_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i0_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i1_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i1_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i1_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i1_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i1_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i1_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_i1_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k0_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k0_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k0_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k0_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k0_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k0_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k0_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k1_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k1_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k1_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k1_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k1_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k1_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_modified_bessel_k1_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtr_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtr_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtr_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtr_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtr_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtr_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtr_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtr_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtr_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtri_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtri_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtri_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtri_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtri_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtri_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtri_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_ndtri_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_polygamma_special_polygamma_n_0_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_polygamma_special_polygamma_n_0_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_polygamma_special_polygamma_n_0_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_polygamma_special_polygamma_n_0_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_polygamma_special_polygamma_n_0_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_polygamma_special_polygamma_n_0_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_polygamma_special_polygamma_n_0_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_polygamma_special_polygamma_n_0_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_polygamma_special_polygamma_n_0_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k0_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k0_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k0_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k0_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k0_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k0_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k0_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k1_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k1_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k1_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k1_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k1_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k1_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_scaled_modified_bessel_k1_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_t_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_t_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_t_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_t_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_t_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_t_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_t_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_t_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_u_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_u_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_u_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_u_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_u_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_u_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_u_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_v_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_v_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_v_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_v_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_v_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_v_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_v_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_v_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_w_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_w_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_w_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_w_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_w_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_w_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_w_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_shifted_chebyshev_polynomial_w_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_spherical_bessel_j0_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_spherical_bessel_j0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_spherical_bessel_j0_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_spherical_bessel_j0_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_spherical_bessel_j0_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_spherical_bessel_j0_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_spherical_bessel_j0_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_spherical_bessel_j0_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_xlog1py_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_xlog1py_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_xlog1py_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_xlog1py_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_xlog1py_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_xlog1py_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_xlog1py_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_xlog1py_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_xlog1py_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_xlog1py_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_zeta_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_zeta_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_zeta_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_zeta_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_zeta_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_zeta_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_zeta_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_special_zeta_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_list_args_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_list_args_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_list_args_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_list_args_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_list_args_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_list_args_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_list_args_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_list_args_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_list_args_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_list_args_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_list_args_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_list_args_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_split_with_sizes_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sqrt_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_square_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_square_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_square_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_square_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_square_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_square_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_square_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_square_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_square_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_square_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_square_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_square_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_squeeze_multiple_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stack_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_mean_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_mean_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_mean_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_mean_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_mean_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_mean_unbiased_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_mean_unbiased_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_mean_unbiased_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_mean_unbiased_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_mean_unbiased_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_mean_unbiased_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_unbiased_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_unbiased_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_unbiased_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_unbiased_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_unbiased_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_std_unbiased_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stft_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stft_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_stft_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sub_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sub_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sub_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sub_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sub_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sub_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sub_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sub_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sub_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sub_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sub_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sub_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_to_size_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_to_size_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_to_size_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_to_size_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_to_size_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_to_size_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_to_size_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_to_size_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_to_size_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_to_size_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_to_size_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_sum_to_size_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_svd_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_svd_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_svd_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_svd_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_svd_lowrank_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_svd_lowrank_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_svd_lowrank_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_svd_lowrank_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_t_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_along_dim_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_along_dim_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_along_dim_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_along_dim_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_along_dim_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_along_dim_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_along_dim_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_along_dim_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_along_dim_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_along_dim_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_along_dim_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_along_dim_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_take_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tan_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tanh_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensor_split_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensor_split_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensor_split_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensor_split_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensor_split_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensor_split_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensor_split_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensor_split_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensor_split_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensor_split_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensor_split_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensor_split_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensordot_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensordot_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensordot_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensordot_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensordot_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tensordot_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tile_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tile_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tile_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tile_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tile_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tile_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tile_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tile_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tile_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tile_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tile_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tile_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_sparse_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_sparse_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_sparse_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_sparse_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_sparse_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_sparse_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_sparse_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_sparse_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_sparse_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_sparse_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_sparse_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_to_sparse_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_topk_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_topk_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_topk_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_topk_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_topk_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_topk_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_topk_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_topk_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_topk_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch__scaled_mm_cuda_float8_e4m3fn, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__efficient_attention_forward_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__efficient_attention_forward_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__efficient_attention_forward_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__flash_attention_forward_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__flash_attention_forward_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__safe_softmax_default_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__safe_softmax_default_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__safe_softmax_default_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__safe_softmax_default_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__safe_softmax_default_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__safe_softmax_default_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__safe_softmax_default_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__safe_softmax_default_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__safe_softmax_default_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_torch_ops_aten__safe_softmax_default_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trace_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_transpose_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapezoid_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapezoid_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapezoid_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapezoid_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapezoid_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapezoid_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapezoid_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapezoid_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapezoid_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapezoid_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapezoid_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapz_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapz_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapz_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapz_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapz_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapz_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapz_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapz_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapz_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapz_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trapz_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triangular_solve_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triangular_solve_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triangular_solve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triangular_solve_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_indices_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_tril_indices_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_indices_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_triu_indices_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_true_divide_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trunc_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trunc_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trunc_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trunc_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trunc_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trunc_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trunc_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trunc_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_trunc_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unbind_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unflatten_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unfold_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_uniform_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_uniform_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_uniform_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_uniform_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_uniform_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_uniform_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_consecutive_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_consecutive_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_consecutive_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_consecutive_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_consecutive_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_consecutive_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_consecutive_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_consecutive_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_consecutive_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_consecutive_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_uint16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_uint32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_uint64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unique_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unravel_index_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unravel_index_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unravel_index_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unravel_index_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unravel_index_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_chunk_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsafe_split_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_unsqueeze_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_mean_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_mean_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_mean_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_mean_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_mean_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_mean_unbiased_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_mean_unbiased_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_mean_unbiased_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_mean_unbiased_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_mean_unbiased_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_mean_unbiased_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_unbiased_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_unbiased_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_unbiased_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_unbiased_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_unbiased_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_var_unbiased_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vdot_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vdot_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vdot_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vdot_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vdot_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vdot_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_complex_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_complex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_complex_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_real_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_as_real_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_copy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_copy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_copy_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_copy_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_copy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_copy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_copy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_copy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_copy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_copy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_copy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_view_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vsplit_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_vstack_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_where_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_xlogy_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_xlogy_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_xlogy_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_xlogy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_xlogy_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_xlogy_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_xlogy_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_xlogy_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_xlogy_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_xlogy_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zero__cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zero__cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zero__cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zero__cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zero__cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zero__cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zero__cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zero__cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zero__cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zero__cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zero__cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zero__cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_bfloat16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_bool, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_complex128, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_complex32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_complex64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_float16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_float64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_int16, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_int32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_int64, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_int8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_nnc_correctness_zeros_like_cuda_uint8, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_H_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_T_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported___getitem___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported___rpow___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported___rsub___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported__batch_norm_with_update_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported__chunk_cat_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported__native_batch_norm_legit_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported__segment_reduce_lengths_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported__segment_reduce_offsets_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported__softmax_backward_data_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported__unsafe_masked_index_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported__unsafe_masked_index_put_accumulate_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported__upsample_bilinear2d_aa_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_acosh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_addbmm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_addcdiv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_addmm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_addmv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_addr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_alias_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_all_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_allclose_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_amax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_amin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_aminmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_angle_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_any_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_arange_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_argmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_argmin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_argsort_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_argwhere_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_as_strided_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_as_strided_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_as_strided_partial_views_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_as_strided_scatter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_asinh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_atanh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_atleast_1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_atleast_2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_atleast_3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_baddbmm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_bernoulli_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_bfloat16_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_block_diag_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_bmm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_broadcast_shapes_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_broadcast_tensors_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_broadcast_to_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_bucketize_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cartesian_prod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cat_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cauchy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cdist_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cdouble_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cfloat_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_chalf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cholesky_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cholesky_inverse_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cholesky_solve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_chunk_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_clamp_max_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_clamp_min_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_clone_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_column_stack_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_combinations_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_complex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_conj_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_conj_physical_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_constant_pad_nd_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_copysign_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_corrcoef_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_count_nonzero_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cov_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cross_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cummax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cummin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cumprod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cumsum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_cumulative_trapezoid_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_deg2rad_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_diag_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_diag_embed_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_diagflat_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_diagonal_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_diagonal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_diagonal_scatter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_diff_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_digamma_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_dist_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_dot_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_dsplit_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_dstack_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_einsum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_empty_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_empty_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_empty_permuted_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_empty_strided_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_equal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_erfinv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_exp2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_expand_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_exponential_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_eye_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_fft2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_fft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_fftn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_fftshift_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_hfft2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_hfft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_hfftn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_ifft2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_ifft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_ifftn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_ifftshift_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_ihfft2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_ihfft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_ihfftn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_irfft2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_irfft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_irfftn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_rfft2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_rfft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fft_rfftn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fill_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_flatten_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_flip_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fliplr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_flipud_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_float_power_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_floor_divide_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_fmin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_frexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_full_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_full_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_gather_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_geometric_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_geqrf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_gradient_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_grid_sampler_2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_heaviside_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_histc_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_hsplit_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_hstack_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_hypot_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_i0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_igamma_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_igammac_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_index_add_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_index_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_index_fill_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_index_put_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_index_reduce_amax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_index_reduce_amin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_index_reduce_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_index_reduce_prod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_index_select_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_inner_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_isclose_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_isfinite_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_isin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_isinf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_isneginf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_isposinf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_isreal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_item_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_jiterator_2inputs_2outputs_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_jiterator_4inputs_with_extra_args_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_jiterator_binary_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_jiterator_binary_return_by_ref_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_jiterator_unary_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_kron_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_kthvalue_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_ldexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_cholesky_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_cholesky_ex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_cond_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_cross_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_det_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_diagonal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_eig_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_eigh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_eigvals_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_eigvalsh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_householder_product_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_inv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_inv_ex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_ldl_factor_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_ldl_factor_ex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_ldl_solve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_lstsq_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_lstsq_grad_oriented_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_lu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_lu_factor_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_lu_factor_ex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_lu_solve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_matrix_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_matrix_power_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_matrix_rank_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_matrix_rank_hermitian_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_multi_dot_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_norm_subgradients_at_zero_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_pinv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_pinv_hermitian_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_pinv_singular_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_qr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_slogdet_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_solve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_solve_ex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_solve_triangular_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_svd_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_svdvals_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_tensorinv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_tensorsolve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_vander_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_vecdot_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linalg_vector_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linspace_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_linspace_tensor_overload_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_log_normal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_log_softmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_log_softmax_with_dtype_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_logaddexp2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_logaddexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_logcumsumexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_logdet_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_logical_and_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_logical_not_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_logical_or_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_logical_xor_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_logit_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_logspace_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_logspace_tensor_overload_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_logsumexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_lu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_lu_solve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_lu_unpack_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_mH_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_mT_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_amax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_amin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_argmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_argmin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_cumprod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_cumsum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_log_softmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_logaddexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_logsumexp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_median_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_normalize_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_prod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_scatter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_select_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_softmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_softmin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_std_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_sum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_masked_var_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_matrix_exp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_max_pool2d_with_indices_backward_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_max_reduction_no_dim_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_max_reduction_with_dim_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_maximum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_median_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_meshgrid_list_of_tensors_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_meshgrid_variadic_tensors_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_min_reduction_no_dim_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_min_reduction_with_dim_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_minimum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_mode_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_movedim_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_msort_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_multinomial_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_mv_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_mvlgamma_mvlgamma_p_1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_mvlgamma_mvlgamma_p_3_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_mvlgamma_mvlgamma_p_5_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nan_to_num_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nanmean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nanmedian_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nanquantile_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nansum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_narrow_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_narrow_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_native_batch_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_native_dropout_backward_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_native_layer_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_new_empty_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_new_empty_strided_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_new_full_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_new_ones_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_new_zeros_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nextafter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_adaptive_avg_pool1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_adaptive_avg_pool2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_adaptive_avg_pool3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_adaptive_max_pool1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_adaptive_max_pool2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_adaptive_max_pool3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_alpha_dropout_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_avg_pool1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_avg_pool2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_avg_pool3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_batch_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_batch_norm_without_cudnn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_bilinear_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_binary_cross_entropy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_celu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_channel_shuffle_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_conv1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_conv2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_conv3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_conv_transpose1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_conv_transpose2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_conv_transpose3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_cosine_embedding_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_cosine_similarity_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_cross_entropy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_ctc_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_dropout2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_dropout3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_dropout_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_elu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_embedding_bag_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_embedding_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_feature_alpha_dropout_with_train_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_feature_alpha_dropout_without_train_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_fractional_max_pool2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_fractional_max_pool3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_gaussian_nll_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_gelu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_glu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_grid_sample_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_group_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_hinge_embedding_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_huber_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_instance_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_interpolate_area_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_interpolate_bicubic_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_interpolate_bilinear_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_interpolate_linear_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_interpolate_nearest-exact_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_interpolate_nearest_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_interpolate_trilinear_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_kl_div_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_l1_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_layer_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_linear_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_local_response_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_logsigmoid_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_margin_ranking_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_max_pool1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_max_pool2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_max_pool3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_max_unpool1d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_max_unpool1d_grad_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_max_unpool2d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_max_unpool2d_grad_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_max_unpool3d_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_max_unpool3d_grad_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_mish_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_mse_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_multi_head_attention_forward_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_multi_margin_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_multilabel_margin_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_multilabel_soft_margin_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_nll_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_normalize_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_pad_circular_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_pad_constant_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_pad_reflect_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_pad_replicate_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_pad_replicate_negative_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_pairwise_distance_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_pdist_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_pixel_shuffle_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_pixel_unshuffle_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_poisson_nll_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_prelu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_rms_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_rrelu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_scaled_dot_product_attention_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_selu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_silu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_smooth_l1_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_soft_margin_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_softmin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_softmin_with_dtype_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_softshrink_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_triplet_margin_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_unfold_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_upsample_bilinear_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nn_functional_upsample_nearest_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nonzero_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_nonzero_static_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_norm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_norm_fro_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_norm_inf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_norm_nuc_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_normal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_normal_in_place_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_normal_number_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_ones_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_ones_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_ormqr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_outer_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_pca_lowrank_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_permute_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_pinverse_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_polar_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_polygamma_polygamma_n_0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_polygamma_polygamma_n_1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_polygamma_polygamma_n_2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_polygamma_polygamma_n_3_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_polygamma_polygamma_n_4_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_positive_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_prod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_put_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_qr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_quantile_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_rad2deg_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_rand_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_randint_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_randint_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_randn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_randn_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_ravel_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_real_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_renorm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_repeat_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_repeat_interleave_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_resize__cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_resize_as__cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_resolve_conj_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_resolve_neg_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_roll_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_rot90_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_round_decimals_0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_round_decimals_3_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_round_decimals_neg_3_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_scalar_tensor_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_scatter_add_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_scatter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_scatter_reduce_amax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_scatter_reduce_amin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_scatter_reduce_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_scatter_reduce_prod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_scatter_reduce_sum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_searchsorted_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_select_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_select_scatter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_sgn_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_signal_windows_bartlett_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_signal_windows_blackman_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_signal_windows_cosine_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_signal_windows_exponential_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_signal_windows_gaussian_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_signal_windows_general_cosine_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_signal_windows_general_hamming_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_signal_windows_hamming_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_signal_windows_hann_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_signal_windows_kaiser_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_signal_windows_nuttall_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_signbit_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_sinc_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_slice_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_slice_scatter_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_softmax_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_softmax_with_dtype_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_sort_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_sparse_mm_reduce_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_sparse_sampled_addmm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_airy_ai_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_bessel_j0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_bessel_j1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_bessel_y0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_bessel_y1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_chebyshev_polynomial_t_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_chebyshev_polynomial_u_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_chebyshev_polynomial_v_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_chebyshev_polynomial_w_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_entr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_erfcx_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_hermite_polynomial_h_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_hermite_polynomial_he_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_i0e_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_i1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_i1e_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_laguerre_polynomial_l_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_legendre_polynomial_p_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_log_ndtr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_modified_bessel_i0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_modified_bessel_i1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_modified_bessel_k0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_modified_bessel_k1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_ndtr_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_ndtri_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_polygamma_special_polygamma_n_0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_scaled_modified_bessel_k0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_scaled_modified_bessel_k1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_shifted_chebyshev_polynomial_t_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_shifted_chebyshev_polynomial_u_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_shifted_chebyshev_polynomial_v_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_shifted_chebyshev_polynomial_w_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_spherical_bessel_j0_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_xlog1py_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_special_zeta_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_split_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_split_list_args_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_split_with_sizes_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_split_with_sizes_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_square_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_squeeze_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_squeeze_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_squeeze_multiple_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_stack_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_std_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_std_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_std_mean_unbiased_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_std_unbiased_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_stft_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_sum_to_size_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_svd_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_svd_lowrank_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_t_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_take_along_dim_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_take_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_tensor_split_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_tensordot_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_tile_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_to_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_to_sparse_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_topk_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_torch_ops_aten__efficient_attention_forward_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_torch_ops_aten__safe_softmax_default_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_trace_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_transpose_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_trapezoid_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_trapz_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_triangular_solve_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_tril_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_triu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_unbind_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_unbind_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_unflatten_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_unfold_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_unfold_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_uniform_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_unique_consecutive_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_unique_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_unsafe_chunk_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_unsafe_split_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_unsqueeze_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_var_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_var_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_var_mean_unbiased_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_var_unbiased_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_vdot_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_view_as_complex_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_view_copy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_vsplit_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_vstack_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_xlogy_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_zero__cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_zeros_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_unsupported_zeros_like_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working___radd___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working___rdiv___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working___rmod___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working___rmul___cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_abs_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_acos_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_add_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_addcmul_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_addmm_decomposed_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_asin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_atan2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_atan_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_bool_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_byte_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_ceil_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_char_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_clamp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_contiguous_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_cos_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_cosh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_div_floor_rounding_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_div_no_rounding_mode_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_div_trunc_rounding_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_double_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_eq_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_erf_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_erfc_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_exp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_expand_as_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_expand_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_expm1_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_float_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_floor_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_fmod_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_ge_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_gt_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_half_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_int_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_isnan_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_le_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_lerp_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_lgamma_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_log10_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_log1p_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_log2_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_log_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_long_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_lt_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_masked_fill_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_max_binary_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_mean_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_min_binary_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_mm_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_mul_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_ne_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_neg_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_nn_functional_hardshrink_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_nn_functional_hardsigmoid_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_nn_functional_hardswish_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_nn_functional_hardtanh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_nn_functional_leaky_relu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_nn_functional_relu6_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_nn_functional_relu_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_nn_functional_softplus_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_nn_functional_softsign_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_nn_functional_tanhshrink_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_nn_functional_threshold_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_permute_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_pow_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_reciprocal_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_remainder_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_reshape_as_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_reshape_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_round_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_rsqrt_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_rsub_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_short_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_sigmoid_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_sign_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_sin_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_sinh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_sqrt_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_sub_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_sum_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_t_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_tan_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_tanh_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_transpose_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_true_divide_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_trunc_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_unsqueeze_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_view_as_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_view_cuda_float32, test/test_jit_fuser_te.py::TestNNCOpInfoCUDA::test_working_where_cuda_float32 2025-07-17T09:41:25.2536721Z 2025-07-17T09:41:25.2536913Z Running test_autoload 1/1 ... [2025-07-17 09:41:24.978135] 2025-07-17T09:41:25.2537208Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:41:25.2537929Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_autoload.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:41:24.978465] 2025-07-17T09:41:28.6005855Z 2025-07-17T09:41:28.6006790Z test_autoload 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_autoload_1.1_8f3cce4c3de6c602_.log 2025-07-17T09:41:28.6007413Z Running 1 items in this shard: test/test_autoload.py::TestDeviceBackendAutoload::test_autoload 2025-07-17T09:41:28.6007682Z 2025-07-17T09:41:28.6007808Z Running test_jit_llga_fuser 1/1 ... [2025-07-17 09:41:28.600486] 2025-07-17T09:41:28.6008075Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:41:28.6009544Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_jit_llga_fuser.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:41:28.600802] 2025-07-17T09:43:07.5277410Z 2025-07-17T09:43:07.5282721Z inductor/test_torchinductor 2/2 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_2.2_58367f3efc5612a9_.log 2025-07-17T09:43:07.5437840Z Running 500 items in this shard: test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_broadcast1, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_broadcast3, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_dense, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_double, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast1_int, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast2_broadcast1, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast2_broadcast3, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast2_dense, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast2_strided, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_broadcast1, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_broadcast2, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_broadcast3, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_double, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_strided, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_broadcast3_transposed, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_dense_broadcast2, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_dense_int, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_dense_transposed, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_double_broadcast1, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_double_broadcast3, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_double_dense, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_broadcast1, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_broadcast2, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_dense, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_double, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_int_int, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_strided_broadcast2, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_strided_dense, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_strided_double, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_strided_int, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_broadcast1, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_broadcast3, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_dense, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_strided, test/inductor/test_torchinductor.py::SweepInputsGPUTest::test_cuda_transposed_transposed, test/inductor/test_torchinductor.py::GPUTests::test__dyn_quant_matmul_4bit_cuda, test/inductor/test_torchinductor.py::GPUTests::test__unsafe_masked_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_adaptive_avg_pool2d2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_adaptive_avg_pool2d_low_prec_cuda, test/inductor/test_torchinductor.py::GPUTests::test_adaptive_avg_pool_errors_with_long_cuda, test/inductor/test_torchinductor.py::GPUTests::test_adaptive_max_pool2d2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_add_complex3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_add_complex_cuda, test/inductor/test_torchinductor.py::GPUTests::test_add_const_int_cuda, test/inductor/test_torchinductor.py::GPUTests::test_adding_tensor_offsets_cuda, test/inductor/test_torchinductor.py::GPUTests::test_addmv_cuda, test/inductor/test_torchinductor.py::GPUTests::test_aliased_buffer_reuse_cuda, test/inductor/test_torchinductor.py::GPUTests::test_aoti_eager_cache_hit_cuda, test/inductor/test_torchinductor.py::GPUTests::test_aoti_eager_dtype_device_layout_cuda, test/inductor/test_torchinductor.py::GPUTests::test_aoti_eager_override_registration_cuda, test/inductor/test_torchinductor.py::GPUTests::test_aoti_eager_support_out_cuda, test/inductor/test_torchinductor.py::GPUTests::test_aoti_eager_support_str_cuda, test/inductor/test_torchinductor.py::GPUTests::test_arange3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_arange4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_arange5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_argmax_argmin3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_argmax_argmin_with_duplicates_cuda, test/inductor/test_torchinductor.py::GPUTests::test_argmax_argmin_with_nan_cuda, test/inductor/test_torchinductor.py::GPUTests::test_argmax_min_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_argmax_to_float_cuda, test/inductor/test_torchinductor.py::GPUTests::test_as_strided_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_assert_alignment_op_name_pass_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d_backward2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d_backward3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d_backward4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool2d_backward_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool3d_backward2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool3d_backward3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool3d_backward_cuda, test/inductor/test_torchinductor.py::GPUTests::test_avg_pool_errors_with_uint_cuda, test/inductor/test_torchinductor.py::GPUTests::test_baddbmm_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bernoulli1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bernoulli2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bfloat16_to_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bitwise3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bitwise_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bmm1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bool_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_broadcast_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_int8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int16_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int32_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int32_int8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int8_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int8_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_int8_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_uint8_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_int_uint8_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_nd_tiling_False_cuda, test/inductor/test_torchinductor.py::GPUTests::test_bucketize_nd_tiling_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_buffer_batch_norm_cuda, test/inductor/test_torchinductor.py::GPUTests::test_buffer_copied_in_graph_cuda, test/inductor/test_torchinductor.py::GPUTests::test_buffer_copied_in_graph_with_different_shapes_cuda, test/inductor/test_torchinductor.py::GPUTests::test_buffer_use_after_remove_cuda, test/inductor/test_torchinductor.py::GPUTests::test_builtins_round_cuda, test/inductor/test_torchinductor.py::GPUTests::test_builtins_round_float_ndigits_pos_cuda, test/inductor/test_torchinductor.py::GPUTests::test_builtins_round_float_ndigits_zero_cuda, test/inductor/test_torchinductor.py::GPUTests::test_builtins_round_int_ndigits_zero_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_extern_kernel_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_inplace_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_negative_dim_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_of_loops_and_extern_kernel_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_unbacked_2d_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_unbacked_legacy_empty_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cat_upcasting_cuda, test/inductor/test_torchinductor.py::GPUTests::test_check_stack_no_cycles_cuda, test/inductor/test_torchinductor.py::GPUTests::test_clamp_cuda, test/inductor/test_torchinductor.py::GPUTests::test_config_option_dont_assume_alignment_cuda, test/inductor/test_torchinductor.py::GPUTests::test_consecutive_split_cumsum_cuda, test/inductor/test_torchinductor.py::GPUTests::test_constant_pad_1d_cuda, test/inductor/test_torchinductor.py::GPUTests::test_constant_pad_2d_cuda, test/inductor/test_torchinductor.py::GPUTests::test_constant_pad_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv3d_channels_last_use_block_ptr_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv_backward_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv_bn_fuse_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv_functional_bn_fuse_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv_inference_heuristics_cuda, test/inductor/test_torchinductor.py::GPUTests::test_conv_with_as_strided_cuda, test/inductor/test_torchinductor.py::GPUTests::test_convolution2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_convolution4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_convolution5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cos_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cumsum_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cumsum_no_mask_cuda, test/inductor/test_torchinductor.py::GPUTests::test_cumsum_zero_dim_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_op_1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_op_2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_op_default_layout_constraint_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_op_fixed_layout_channels_last_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_op_fixed_layout_sequential_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_scan_op_compiled_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_scan_op_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_scan_op_multi_input_cuda, test/inductor/test_torchinductor.py::GPUTests::test_custom_scan_would_split_cuda, test/inductor/test_torchinductor.py::GPUTests::test_deterministic_codegen_cuda, test/inductor/test_torchinductor.py::GPUTests::test_diagonal_copy_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dist_bf16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dist_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div6_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div9_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div_prim_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div_softmax_symfloat_cuda, test/inductor/test_torchinductor.py::GPUTests::test_div_zero_dim_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dropout_trivial_0_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtype_sympy_expr_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_bfloat16_bfloat16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_bfloat16_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_bfloat16_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float16_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_int8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float32_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float64_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_float64_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int16_int8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int32_int8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_int8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int64_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_bfloat16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_int16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_int8_int64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_bfloat16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_float16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_dtypeview_uint8_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_embedding_cuda, test/inductor/test_torchinductor.py::GPUTests::test_empty1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_empty2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_erfc_cuda, test/inductor/test_torchinductor.py::GPUTests::test_erfinv_cuda, test/inductor/test_torchinductor.py::GPUTests::test_exact_stride_cuda, test/inductor/test_torchinductor.py::GPUTests::test_expm1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fallback_mutable_op_list_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fallback_mutable_op_no_mutated_tensors_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fft_real_input_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fill2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_flip_cat_cuda, test/inductor/test_torchinductor.py::GPUTests::test_flip_cuda, test/inductor/test_torchinductor.py::GPUTests::test_float_index_expression_cuda, test/inductor/test_torchinductor.py::GPUTests::test_float_repr_dynamic_shapes_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fmin_fmax_cuda, test/inductor/test_torchinductor.py::GPUTests::test_forced_buffer_realize_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fractional_max_pool2d5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_fuse_tiled_cuda, test/inductor/test_torchinductor.py::GPUTests::test_gather2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_gather_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_getitem_cuda, test/inductor/test_torchinductor.py::GPUTests::test_graph_partition_arange1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_graph_partition_constant_tensor2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_graph_partition_misaligned_input_cuda, test/inductor/test_torchinductor.py::GPUTests::test_graph_partition_no_inputs_cuda, test/inductor/test_torchinductor.py::GPUTests::test_graph_partition_unbacked_symint_as_output_cuda, test/inductor/test_torchinductor.py::GPUTests::test_hardsigmoid_cuda, test/inductor/test_torchinductor.py::GPUTests::test_hardswish_cuda, test/inductor/test_torchinductor.py::GPUTests::test_horizonal_fusion2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_dynamic_shapes_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_propagation_abs_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_propagation_device_assert_masked_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_propagation_nested_indirect_indexing_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_put1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_put2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_put4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_put_deterministic_fallback_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_put_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_put_reinplace_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_remainder_cuda, test/inductor/test_torchinductor.py::GPUTests::test_index_select_cuda, test/inductor/test_torchinductor.py::GPUTests::test_inductor_multiple_specializations_cuda, test/inductor/test_torchinductor.py::GPUTests::test_inf_cuda, test/inductor/test_torchinductor.py::GPUTests::test_inner_fn_str_and_stride_cuda, test/inductor/test_torchinductor.py::GPUTests::test_inplace_activations_cuda, test/inductor/test_torchinductor.py::GPUTests::test_inplace_add_cuda, test/inductor/test_torchinductor.py::GPUTests::test_input_mutation1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_input_mutation3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_input_mutation4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_input_mutation5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_insignificant_strides_cuda, test/inductor/test_torchinductor.py::GPUTests::test_isinf_cuda, test/inductor/test_torchinductor.py::GPUTests::test_issue102546_cuda, test/inductor/test_torchinductor.py::GPUTests::test_kernel_names_cuda, test/inductor/test_torchinductor.py::GPUTests::test_kwargs_cuda, test/inductor/test_torchinductor.py::GPUTests::test_l1_loss_cuda, test/inductor/test_torchinductor.py::GPUTests::test_large_broadcast_reduction_cuda, test/inductor/test_torchinductor.py::GPUTests::test_large_grid_use_block_ptr_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_large_pointwise_cuda, test/inductor/test_torchinductor.py::GPUTests::test_large_strided_reduction_cuda, test/inductor/test_torchinductor.py::GPUTests::test_large_tensor_reduction_cuda, test/inductor/test_torchinductor.py::GPUTests::test_leaky_relu_cuda, test/inductor/test_torchinductor.py::GPUTests::test_lgamma_cuda, test/inductor/test_torchinductor.py::GPUTests::test_like_rands3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linear1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linear2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linear_mixed_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linspace1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linspace2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_linspace4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_log_fp64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_log_softmax_cuda, test/inductor/test_torchinductor.py::GPUTests::test_logaddexp_cuda, test/inductor/test_torchinductor.py::GPUTests::test_logcumsumexp_cuda, test/inductor/test_torchinductor.py::GPUTests::test_logcumsumexp_zero_dim_cuda, test/inductor/test_torchinductor.py::GPUTests::test_low_memory_max_pool_dilation_2_dim_2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_low_memory_max_pool_dilation_2_dim_3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_masked_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_matmul_layer_norm_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_min_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_pool2d3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_pool2d5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_pool2d_with_indices_backward2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_pool2d_with_indices_backward3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_pool2d_with_indices_backward4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_max_pool2d_with_indices_backward6_cuda, test/inductor/test_torchinductor.py::GPUTests::test_min_max_reduction_nan_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mix_device_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mixed_mm2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mixed_mm3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mixed_mm_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mm_mixed_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mm_views_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mul_index_expr_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mul_softmax_symfloat_cuda, test/inductor/test_torchinductor.py::GPUTests::test_multi_gpu_device_cuda, test/inductor/test_torchinductor.py::GPUTests::test_multi_gpu_recompile_on_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_multilayer_var_cuda, test/inductor/test_torchinductor.py::GPUTests::test_multilayer_var_lowp_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mutable_custom_op_fixed_layout2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_mutable_custom_op_fixed_layout_cuda, test/inductor/test_torchinductor.py::GPUTests::test_nan_to_num_cuda, test/inductor/test_torchinductor.py::GPUTests::test_neg_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_neg_max_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_nll_loss_backward_cuda, test/inductor/test_torchinductor.py::GPUTests::test_nll_loss_forward_cuda, test/inductor/test_torchinductor.py::GPUTests::test_no_mega_fusion_during_lowering_cuda, test/inductor/test_torchinductor.py::GPUTests::test_no_op_reduction_cuda, test/inductor/test_torchinductor.py::GPUTests::test_no_specization_over_symbolic_value_cuda, test/inductor/test_torchinductor.py::GPUTests::test_nonzero_unbacked_refinement_cuda, test/inductor/test_torchinductor.py::GPUTests::test_norm_constant_overflow_cuda, test/inductor/test_torchinductor.py::GPUTests::test_output_strides_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pattern_matcher_multi_user_cuda, test/inductor/test_torchinductor.py::GPUTests::test_permute1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_permute2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_bessel_j0_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_bessel_j1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_chebyshev_polynomial_u_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_chebyshev_polynomial_v_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_chebyshev_polynomial_w_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_digamma_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_entr_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_erfc_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_expit_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_expm1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_gammainc_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_gammaincc_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_hermite_polynomial_he_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_i0_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_i0e_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_i1e_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_logit_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_modified_bessel_k0_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_ndtr_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_polygamma_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_scaled_modified_bessel_k1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_shifted_chebyshev_polynomial_t_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_shifted_chebyshev_polynomial_u_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_sinc_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_xlog1py_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pointwise_xlogy_cuda, test/inductor/test_torchinductor.py::GPUTests::test_polar_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pow1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pow2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pow3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pow_int_cuda, test/inductor/test_torchinductor.py::GPUTests::test_pow_symfloat_cuda, test/inductor/test_torchinductor.py::GPUTests::test_prepare_softmax_with_fast_math_cuda, test/inductor/test_torchinductor.py::GPUTests::test_profiler_mark_wrapper_call_cuda, test/inductor/test_torchinductor.py::GPUTests::test_randint_cuda, test/inductor/test_torchinductor.py::GPUTests::test_randint_distribution_cuda, test/inductor/test_torchinductor.py::GPUTests::test_randint_int64_mod_cuda, test/inductor/test_torchinductor.py::GPUTests::test_randn_generator_cuda, test/inductor/test_torchinductor.py::GPUTests::test_randn_like_empty_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reduction1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reduction2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reduction3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reduction5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reduction_config_limit_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reflection_pad2d_backward_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reinterpret_dtypeview_cuda, test/inductor/test_torchinductor.py::GPUTests::test_relu_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_no_ops_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_default_cuda, test/inductor/test_torchinductor.py::GPUTests::test_remove_noop_view_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_repeat_interleave_2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_replication_pad_errors_with_bool_cuda, test/inductor/test_torchinductor.py::GPUTests::test_require_stride_expanded_cuda, test/inductor/test_torchinductor.py::GPUTests::test_resize_as_cuda, test/inductor/test_torchinductor.py::GPUTests::test_resize_cuda, test/inductor/test_torchinductor.py::GPUTests::test_reuse_buffers_with_aliasing_cuda, test/inductor/test_torchinductor.py::GPUTests::test_roll_cuda, test/inductor/test_torchinductor.py::GPUTests::test_rsqrt_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scalar_output_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scaled_dot_product_attention_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter_add1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter_add3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_scatter_reduce1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sdpa_prefer_nd_tiling_True_use_block_ptr_False_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sdpa_unaligned_mask_freezing_cuda, test/inductor/test_torchinductor.py::GPUTests::test_searchsorted_cuda, test/inductor/test_torchinductor.py::GPUTests::test_select_scatter_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sgn_extremal_cuda, test/inductor/test_torchinductor.py::GPUTests::test_shape_padding_cuda, test/inductor/test_torchinductor.py::GPUTests::test_should_pad_bench_for_bmm_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sigmoid_cuda, test/inductor/test_torchinductor.py::GPUTests::test_signbit_cuda, test/inductor/test_torchinductor.py::GPUTests::test_silu_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sin_cuda, test/inductor/test_torchinductor.py::GPUTests::test_single_elem_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_mutation1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_mutation2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_mutation3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_scatter2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_scatter4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_scatter5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_slice_scatter_reinplace_cuda, test/inductor/test_torchinductor.py::GPUTests::test_softmax_backward_data_cuda, test/inductor/test_torchinductor.py::GPUTests::test_softmax_one_kernel_loop_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sort_bool_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sort_transpose_cuda, test/inductor/test_torchinductor.py::GPUTests::test_special_polygamma_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_cumprod_low_prec_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_cumsum_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_cumsum_index_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_cumsum_low_prec_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_reduction_dynamic_shape_cuda, test/inductor/test_torchinductor.py::GPUTests::test_split_with_integer_cuda, test/inductor/test_torchinductor.py::GPUTests::test_squeeze2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_squeeze_varargs_cuda, test/inductor/test_torchinductor.py::GPUTests::test_std_cuda, test/inductor/test_torchinductor.py::GPUTests::test_stride_preservation_with_stride_modifying_fx_pass_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sum1_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sum2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sum3_cuda, test/inductor/test_torchinductor.py::GPUTests::test_sum5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_tensor2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_tensor_index_slice_cuda, test/inductor/test_torchinductor.py::GPUTests::test_tmp_not_defined_issue1_use_block_ptr_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_tmp_not_defined_issue2_cuda, test/inductor/test_torchinductor.py::GPUTests::test_to_device_constant_cuda, test/inductor/test_torchinductor.py::GPUTests::test_to_device_cuda, test/inductor/test_torchinductor.py::GPUTests::test_to_dtype_cuda, test/inductor/test_torchinductor.py::GPUTests::test_to_memory_format_cuda, test/inductor/test_torchinductor.py::GPUTests::test_transposed_propagates_cuda, test/inductor/test_torchinductor.py::GPUTests::test_triu_cuda, test/inductor/test_torchinductor.py::GPUTests::test_uint4x2_mixed_mm_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unbacked_floordiv_simplify_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unbacked_floordiv_simplify_errors_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unbind_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unfold_zero_dimension_tensor_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_bfloat16_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_float32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_float64_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_int32_cuda, test/inductor/test_torchinductor.py::GPUTests::test_unspec_inputs_uint8_cuda, test/inductor/test_torchinductor.py::GPUTests::test_upsample_bilinear2d_a_cuda, test/inductor/test_torchinductor.py::GPUTests::test_upsample_bilinear2d_b_cuda, test/inductor/test_torchinductor.py::GPUTests::test_upsample_nearest1d_cuda, test/inductor/test_torchinductor.py::GPUTests::test_upsample_nearest2d_cuda, test/inductor/test_torchinductor.py::GPUTests::test_upsample_nearest3d_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_correction_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_False_cuda, test/inductor/test_torchinductor.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_torchinductor.py::GPUTests::test_vectorized_ops_masked_cuda, test/inductor/test_torchinductor.py::GPUTests::test_vectorized_ops_masked_var_novec_cuda, test/inductor/test_torchinductor.py::GPUTests::test_view_as_complex_cuda, test/inductor/test_torchinductor.py::GPUTests::test_view_detach_cuda, test/inductor/test_torchinductor.py::GPUTests::test_view_on_aliased_cuda, test/inductor/test_torchinductor.py::GPUTests::test_view_uint8_through_differing_bitwidths_cuda, test/inductor/test_torchinductor.py::GPUTests::test_views4_cuda, test/inductor/test_torchinductor.py::GPUTests::test_views5_cuda, test/inductor/test_torchinductor.py::GPUTests::test_views6_cuda, test/inductor/test_torchinductor.py::GPUTests::test_views7_cuda, test/inductor/test_torchinductor.py::GPUTests::test_weight_norm_bwd_cuda, test/inductor/test_torchinductor.py::GPUTests::test_xblock_divides_xnumel_cuda, test/inductor/test_torchinductor.py::GPUTests::test_zero_dim_reductions_cuda, test/inductor/test_torchinductor.py::GPUTests::test_zero_element_mutation_cuda, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_cant_optimize_compute, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_ctr_not_moved_to_cuda_when_used_in_index_put, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_donated_buffer_inplace, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_evict_last_non_coalesced_loads_block_ptr, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_graph_partition, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_graph_partition_condition_op, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_graph_partition_dynamic_scalar_inputs, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_graph_partition_fused_scheduler_node, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_graph_partition_item, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_graph_partition_symint, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_graph_partition_unbacked_symint, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_grouped_mm, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_has_constant_mask_block_multiple_False_ynumel_exceed_ygrid_size_False, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_has_constant_mask_block_multiple_True_ynumel_exceed_ygrid_size_False, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_has_constant_mask_block_multiple_True_ynumel_exceed_ygrid_size_True, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_inductor_detach_view_backend_inductor, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_inductor_sequence_nr, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_not_materialize_pointwise_reduction, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_numpy_on_gpu, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_optimize_compute, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_optimize_indexing_assert, test/inductor/test_torchinductor.py::TritonCodeGenTests::test_optimize_indexing_dtype_with_constraint, test/inductor/test_torchinductor.py::RNNTest::test_rnn_compile_safe, test/inductor/test_torchinductor.py::NanCheckerTest::test_nan_checker_pass 2025-07-17T09:43:07.5587796Z 2025-07-17T09:43:07.5588041Z Running dynamo/test_cudagraphs 1/1 ... [2025-07-17 09:43:07.528135] 2025-07-17T09:43:07.5588465Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:43:07.5589727Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_cudagraphs.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:43:07.528479] 2025-07-17T09:43:15.1115629Z 2025-07-17T09:43:15.1116639Z dynamo/test_cudagraphs 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_cudagraphs_1.1_db1624b08d903645_.log 2025-07-17T09:43:15.1118436Z Running 8 items in this shard: test/dynamo/test_cudagraphs.py::TestAotCudagraphs::test_basic, test/dynamo/test_cudagraphs.py::TestAotCudagraphs::test_dead_fill, test/dynamo/test_cudagraphs.py::TestAotCudagraphs::test_dtoh, test/dynamo/test_cudagraphs.py::TestAotCudagraphs::test_factory, test/dynamo/test_cudagraphs.py::TestAotCudagraphs::test_htod, test/dynamo/test_cudagraphs.py::TestAotCudagraphs::test_mutate_constant, test/dynamo/test_cudagraphs.py::TestAotCudagraphs::test_mutate_input, test/dynamo/test_cudagraphs.py::TestAotCudagraphs::test_mutated_metadata 2025-07-17T09:43:15.1120547Z 2025-07-17T09:43:15.1120743Z Running dynamo/test_cudagraphs_expandable_segments 1/1 ... [2025-07-17 09:43:15.111375] 2025-07-17T09:43:15.1121075Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:43:15.1121826Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_cudagraphs_expandable_segments.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:43:15.111694] 2025-07-17T09:43:15.9033166Z 2025-07-17T09:43:15.9034296Z test_jit_llga_fuser 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_jit_llga_fuser_1.1_52390a1d05d96657_.log 2025-07-17T09:43:15.9052872Z Running 107 items in this shard: test/test_jit_llga_fuser.py::TestEnableDisableLlgaFuser::test_context_manager, test/test_jit_llga_fuser.py::TestDynamoAOT::test_dynamo_aot_ts_onednn, test/test_jit_llga_fuser.py::TestModel::test_vision_alexnet_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_alexnet_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_densenet121_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_densenet121_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_densenet161_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_densenet161_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_densenet169_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_densenet169_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_densenet201_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_densenet201_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b0_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b0_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b1_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b1_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b2_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b2_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b3_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b3_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b4_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b4_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b5_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b5_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b6_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b6_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b7_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_efficientnet_b7_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_googlenet_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_googlenet_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_mnasnet1_0_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_mnasnet1_0_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_mobilenet_v2_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_mobilenet_v2_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_mobilenet_v3_large_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_mobilenet_v3_large_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_regnet_y_400mf_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_regnet_y_400mf_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_resnet50_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_resnet50_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_resnext101_32x8d_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_resnext101_32x8d_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_resnext50_32x4d_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_resnext50_32x4d_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_shufflenet_v2_x1_0_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_shufflenet_v2_x1_0_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_squeezenet1_0_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_squeezenet1_0_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_vgg16_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_vgg16_float32, test/test_jit_llga_fuser.py::TestModel::test_vision_wide_resnet50_2_bfloat16, test/test_jit_llga_fuser.py::TestModel::test_vision_wide_resnet50_2_float32, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_bn2d_eltwise_cuda_bfloat16, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_bn2d_eltwise_cuda_float32, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_conv2d_bn_cuda_bfloat16, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_conv2d_bn_cuda_float32, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_conv2d_bn_relu_cuda_bfloat16, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_conv2d_bn_relu_cuda_float32, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_conv2d_clamp_cuda_bfloat16, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_conv2d_clamp_cuda_float32, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_conv2d_eltwise_cuda_bfloat16, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_conv2d_eltwise_cuda_float32, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_conv2d_silu_cuda_bfloat16, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_conv2d_silu_cuda_float32, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_conv2d_sum_cuda_bfloat16, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_conv2d_sum_cuda_float32, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_ensure_tensor_is_rewrapped_cuda_bfloat16, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_ensure_tensor_is_rewrapped_cuda_float32, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_linear_eltwise_cuda_bfloat16, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_linear_eltwise_cuda_float32, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_rewrap_tensor_input_to_pytorch_cuda_bfloat16, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_rewrap_tensor_input_to_pytorch_cuda_float32, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_wildcard_cuda_bfloat16, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_wildcard_cuda_float32, test/test_jit_llga_fuser.py::TestFusionPatternCUDA::test_wildcard_unsupported_dtype_cuda_int32, test/test_jit_llga_fuser.py::TestOpCUDA::test_add_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_add_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_add_scalar_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_add_scalar_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_addmm_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_addmm_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_avg_pool2d_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_avg_pool2d_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_bn2d_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_bn2d_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_cat_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_cat_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_conv2d_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_conv2d_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_eltwise_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_eltwise_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_identity_binary_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_identity_binary_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_layer_norm_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_layer_norm_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_linear_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_linear_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_max_pool2d_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_max_pool2d_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_mul_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_mul_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_softmax_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_softmax_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_typecheck_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_typecheck_cuda_float32, test/test_jit_llga_fuser.py::TestOpCUDA::test_variable_kernel_avg_pool2d_cuda_bfloat16, test/test_jit_llga_fuser.py::TestOpCUDA::test_variable_kernel_avg_pool2d_cuda_float32 2025-07-17T09:43:15.9070576Z 2025-07-17T09:43:15.9070717Z Running dynamo/test_debug_utils 1/1 ... [2025-07-17 09:43:15.903266] 2025-07-17T09:43:15.9071002Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:43:15.9071681Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_debug_utils.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:43:15.903601] 2025-07-17T09:43:22.4398954Z 2025-07-17T09:43:22.4399983Z dynamo/test_cudagraphs_expandable_segments 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_cudagraphs_expandable_segments_1.1_43a60042626df62b_.log 2025-07-17T09:43:22.4402257Z Running 8 items in this shard: test/dynamo/test_cudagraphs_expandable_segments.py::TestAotCudagraphs::test_basic, test/dynamo/test_cudagraphs_expandable_segments.py::TestAotCudagraphs::test_dead_fill, test/dynamo/test_cudagraphs_expandable_segments.py::TestAotCudagraphs::test_dtoh, test/dynamo/test_cudagraphs_expandable_segments.py::TestAotCudagraphs::test_factory, test/dynamo/test_cudagraphs_expandable_segments.py::TestAotCudagraphs::test_htod, test/dynamo/test_cudagraphs_expandable_segments.py::TestAotCudagraphs::test_mutate_constant, test/dynamo/test_cudagraphs_expandable_segments.py::TestAotCudagraphs::test_mutate_input, test/dynamo/test_cudagraphs_expandable_segments.py::TestAotCudagraphs::test_mutated_metadata 2025-07-17T09:43:22.4404071Z 2025-07-17T09:43:22.4404235Z Running dynamo/test_decorators 1/1 ... [2025-07-17 09:43:22.439648] 2025-07-17T09:43:22.4404517Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:43:22.4405193Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_decorators.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:43:22.439957] 2025-07-17T09:43:24.2350349Z 2025-07-17T09:43:24.2352016Z dynamo/test_debug_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_debug_utils_1.1_cde3a59a1ed2f08b_.log 2025-07-17T09:43:24.2353492Z Running 4 items in this shard: test/dynamo/test_debug_utils.py::TestDebugUtilsCUDA::test_cast_model_to_fp64_dtype_args_cuda, test/dynamo/test_debug_utils.py::TestDebugUtilsCUDA::test_generate_env_vars_string_cuda, test/dynamo/test_debug_utils.py::TestDebugUtilsDeviceCUDA::test_aot_graph_parser_cuda, test/dynamo/test_debug_utils.py::TestDebugUtilsDeviceCUDA::test_sym_aot_graph_parser_cuda 2025-07-17T09:43:24.2354589Z 2025-07-17T09:43:24.2354756Z Running dynamo/test_deque_reconstruct 1/1 ... [2025-07-17 09:43:24.234836] 2025-07-17T09:43:24.2355051Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:43:24.2355742Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_deque_reconstruct.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:43:24.235116] 2025-07-17T09:43:28.0077041Z 2025-07-17T09:43:28.0078712Z dynamo/test_deque_reconstruct 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_deque_reconstruct_1.1_ae9bac7af823876a_.log 2025-07-17T09:43:28.0080023Z Running 3 items in this shard: test/dynamo/test_deque_reconstruct.py::TestDequeReconstruct::test_deque_reconstruct_in_globals, test/dynamo/test_deque_reconstruct.py::TestDequeReconstruct::test_deque_reconstruct_not_in_globals, test/dynamo/test_deque_reconstruct.py::TestDequeReconstruct::test_deque_reconstruct_shallows_globals 2025-07-17T09:43:28.0080869Z 2025-07-17T09:43:28.0081014Z Running dynamo/test_deviceguard 1/1 ... [2025-07-17 09:43:28.007487] 2025-07-17T09:43:28.0081297Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:43:28.0081996Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_deviceguard.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:43:28.007793] 2025-07-17T09:43:31.7792472Z 2025-07-17T09:43:31.7793480Z dynamo/test_deviceguard 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_deviceguard_1.1_ec4168242f0ded12_.log 2025-07-17T09:43:31.7795577Z Running 4 items in this shard: test/dynamo/test_deviceguard.py::TestDeviceGuard::test_device_guard, test/dynamo/test_deviceguard.py::TestDeviceGuard::test_device_guard_no_index, test/dynamo/test_deviceguard.py::TestCUDADeviceGuard::test_device_guard, test/dynamo/test_deviceguard.py::TestCUDADeviceGuard::test_device_guard_no_index 2025-07-17T09:43:31.7796413Z 2025-07-17T09:43:31.7796553Z Running dynamo/test_dicts 1/1 ... [2025-07-17 09:43:31.779053] 2025-07-17T09:43:31.7796818Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:43:31.7797479Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_dicts.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:43:31.779359] 2025-07-17T09:43:38.1356278Z 2025-07-17T09:43:38.1357044Z dynamo/test_decorators 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_decorators_1.1_2b5471ce67402323_.log 2025-07-17T09:43:38.1370736Z Running 62 items in this shard: test/dynamo/test_decorators.py::DecoratorTests::test_allow_in_graph, test/dynamo/test_decorators.py::DecoratorTests::test_allow_in_graph_no_id_reuse, test/dynamo/test_decorators.py::DecoratorTests::test_assume_constant_result_on_computation_with_graph_input, test/dynamo/test_decorators.py::DecoratorTests::test_assume_constant_result_on_user_defined_fn, test/dynamo/test_decorators.py::DecoratorTests::test_class_methods, test/dynamo/test_decorators.py::DecoratorTests::test_disable_for_custom_op, test/dynamo/test_decorators.py::DecoratorTests::test_disable_ignores_outer_wraps, test/dynamo/test_decorators.py::DecoratorTests::test_disable_nn_module_with_class_decorator, test/dynamo/test_decorators.py::DecoratorTests::test_disable_nn_modules_forward_hook, test/dynamo/test_decorators.py::DecoratorTests::test_disable_optimize, test/dynamo/test_decorators.py::DecoratorTests::test_disable_recursive_false, test/dynamo/test_decorators.py::DecoratorTests::test_disable_recursive_false_weird, test/dynamo/test_decorators.py::DecoratorTests::test_disallow_in_graph, test/dynamo/test_decorators.py::DecoratorTests::test_dont_skip_tracing, test/dynamo/test_decorators.py::DecoratorTests::test_graph_break, test/dynamo/test_decorators.py::DecoratorTests::test_incorrect_usage_disallow_in_graph, test/dynamo/test_decorators.py::DecoratorTests::test_mark_static_address_guarded, test/dynamo/test_decorators.py::DecoratorTests::test_mark_static_address_unguarded, test/dynamo/test_decorators.py::DecoratorTests::test_mark_static_nn_module, test/dynamo/test_decorators.py::DecoratorTests::test_nested_compile_fullgraph, test/dynamo/test_decorators.py::DecoratorTests::test_nested_disable_decorator, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_newly_constructed_trace_register_constant_type_error, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_captured_external_tensor, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_custom_class_error, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_inside_compiled_function, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_inside_compiled_function_error, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_inside_compiled_function_kwarg, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_nested_custom_class, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_nested_custom_class_error, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_newly_constructed_custom_class_with_side_effects, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_newly_constructed_dict_with_side_effects, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_no_action_at_a_distance, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_object_in_context_error, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_on_method, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_pre_existing_custom_class, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_pre_existing_custom_class_with_side_effects, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_pre_existing_dict, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_pre_existing_dict_with_side_effects, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_pre_existing_register_constant_type_guard, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_tensor_args, test/dynamo/test_decorators.py::DecoratorTests::test_nonstrict_trace_tuple_and_sym_int_output, test/dynamo/test_decorators.py::DecoratorTests::test_patch_dynamo_config_errors, test/dynamo/test_decorators.py::DecoratorTests::test_set_fullgraph, test/dynamo/test_decorators.py::DecoratorTests::test_set_fullgraph_error, test/dynamo/test_decorators.py::DecoratorTests::test_set_fullgraph_export, test/dynamo/test_decorators.py::DecoratorTests::test_set_fullgraph_nested, test/dynamo/test_decorators.py::DecoratorTests::test_set_fullgraph_nested_deep, test/dynamo/test_decorators.py::DecoratorTests::test_set_fullgraph_nested_with_skip, test/dynamo/test_decorators.py::DecoratorTests::test_set_stance_aot_eager_then_compile, test/dynamo/test_decorators.py::DecoratorTests::test_set_stance_eager_on_recompile, test/dynamo/test_decorators.py::DecoratorTests::test_set_stance_eager_then_compile, test/dynamo/test_decorators.py::DecoratorTests::test_set_stance_eager_then_compile_with_graph_break, test/dynamo/test_decorators.py::DecoratorTests::test_set_stance_fail_on_recompile, test/dynamo/test_decorators.py::DecoratorTests::test_set_stance_fail_on_recompile_with_disable, test/dynamo/test_decorators.py::DecoratorTests::test_set_stance_forbid_in_graph, test/dynamo/test_decorators.py::DecoratorTests::test_set_stance_force_backend, test/dynamo/test_decorators.py::DecoratorTests::test_set_stance_force_backend_with_disable, test/dynamo/test_decorators.py::DecoratorTests::test_set_stance_force_eager, test/dynamo/test_decorators.py::DecoratorTests::test_skip_frame, test/dynamo/test_decorators.py::DecoratorTests::test_substitute_in_graph, test/dynamo/test_decorators.py::DecoratorTests::test_torch_guards_stack_frame_register_inlining_disable, test/dynamo/test_decorators.py::DecoratorTests::test_torch_guards_stack_frame_register_inlining_partially_disable 2025-07-17T09:43:38.1383496Z 2025-07-17T09:43:38.1383642Z Running dynamo/test_dynamic_shapes 1/3 ... [2025-07-17 09:43:38.135452] 2025-07-17T09:43:38.1383927Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:43:38.1384592Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'dynamo/test_dynamic_shapes.py', '-m', 'not serial', '--shard-id=1', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:43:38.135761] 2025-07-17T09:43:45.0182037Z 2025-07-17T09:43:45.0182740Z dynamo/test_dicts 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_dicts_1.1_487ab442866b1d0a_.log 2025-07-17T09:43:45.0199587Z Running 105 items in this shard: test/dynamo/test_dicts.py::DictTests::test_builtin_ior_, test/dynamo/test_dicts.py::DictTests::test_builtin_or_with_diff_keys, test/dynamo/test_dicts.py::DictTests::test_builtin_or_with_invalid_types, test/dynamo/test_dicts.py::DictTests::test_builtin_or_with_same_keys, test/dynamo/test_dicts.py::DictTests::test_contains_dunder_dict, test/dynamo/test_dicts.py::DictTests::test_contains_module_dunder_dict, test/dynamo/test_dicts.py::DictTests::test_custom_iter_dict, test/dynamo/test_dicts.py::DictTests::test_custom_keys_iter_dict, test/dynamo/test_dicts.py::DictTests::test_dict_construction_from_mapping_proxy, test/dynamo/test_dicts.py::DictTests::test_dict_contains, test/dynamo/test_dicts.py::DictTests::test_dict_copy_alias, test/dynamo/test_dicts.py::DictTests::test_dict_guard_on_keys_order, test/dynamo/test_dicts.py::DictTests::test_dict_guard_on_keys_order2, test/dynamo/test_dicts.py::DictTests::test_dict_iter, test/dynamo/test_dicts.py::DictTests::test_dict_keys_binop_op_and_, test/dynamo/test_dicts.py::DictTests::test_dict_keys_binop_op_or_, test/dynamo/test_dicts.py::DictTests::test_dict_keys_binop_op_sub, test/dynamo/test_dicts.py::DictTests::test_dict_keys_binop_op_xor, test/dynamo/test_dicts.py::DictTests::test_dict_keys_inplace_binop_op_iand, test/dynamo/test_dicts.py::DictTests::test_dict_keys_inplace_binop_op_ior, test/dynamo/test_dicts.py::DictTests::test_dict_keys_inplace_binop_op_isub, test/dynamo/test_dicts.py::DictTests::test_dict_keys_inplace_binop_op_ixor, test/dynamo/test_dicts.py::DictTests::test_dict_list_values, test/dynamo/test_dicts.py::DictTests::test_dict_mutation_side_effect, test/dynamo/test_dicts.py::DictTests::test_dict_namedtuple, test/dynamo/test_dicts.py::DictTests::test_dict_order_keys, test/dynamo/test_dicts.py::DictTests::test_dict_order_keys_modules, test/dynamo/test_dicts.py::DictTests::test_dict_order_keys_tensors, test/dynamo/test_dicts.py::DictTests::test_dict_reconstruct_keeps_original_order, test/dynamo/test_dicts.py::DictTests::test_dict_subclass_contains, test/dynamo/test_dicts.py::DictTests::test_dict_subclass_get_method, test/dynamo/test_dicts.py::DictTests::test_dict_subclass_initialization_in_graph, test/dynamo/test_dicts.py::DictTests::test_dict_subclass_instantiation, test/dynamo/test_dicts.py::DictTests::test_dict_subclass_instantiation_return, test/dynamo/test_dicts.py::DictTests::test_dict_subclass_local_mutation, test/dynamo/test_dicts.py::DictTests::test_dict_subclass_local_with_non_dict_method, test/dynamo/test_dicts.py::DictTests::test_dict_subclass_methods_fallback_mutation, test/dynamo/test_dicts.py::DictTests::test_dict_subclass_methods_fallback_readonly, test/dynamo/test_dicts.py::DictTests::test_dict_subclass_setitem, test/dynamo/test_dicts.py::DictTests::test_dict_tag_guard, test/dynamo/test_dicts.py::DictTests::test_empty_dict_recompilation, test/dynamo/test_dicts.py::DictTests::test_fn_id, test/dynamo/test_dicts.py::DictTests::test_items_type, test/dynamo/test_dicts.py::DictTests::test_lazy_key_guarding, test/dynamo/test_dicts.py::DictTests::test_lazy_key_non_const_guarding, test/dynamo/test_dicts.py::DictTests::test_mapping_proxy_existing, test/dynamo/test_dicts.py::DictTests::test_mapping_proxy_existing_local_mutation, test/dynamo/test_dicts.py::DictTests::test_mapping_proxy_existing_mutation, test/dynamo/test_dicts.py::DictTests::test_mapping_proxy_for_local, test/dynamo/test_dicts.py::DictTests::test_mapping_proxy_for_nonlocal, test/dynamo/test_dicts.py::DictTests::test_move_to_end, test/dynamo/test_dicts.py::DictTests::test_newly_constructed_default_dict, test/dynamo/test_dicts.py::DictTests::test_ordered_dict_reordered_keys, test/dynamo/test_dicts.py::DictTests::test_ordered_dict_subclass_reordered_keys, test/dynamo/test_dicts.py::DictTests::test_overridden_get_item, test/dynamo/test_dicts.py::DictTests::test_udf_dict_reconstruction, test/dynamo/test_dicts.py::DictTests::test_update_dunder_dict, test/dynamo/test_dicts.py::DictTests::test_update_module_dunder_dict, test/dynamo/test_dicts.py::DictTests::test_weakref_dict, test/dynamo/test_dicts.py::DictGuardTests::test_cmp_eq, test/dynamo/test_dicts.py::DictGuardTests::test_cmp_ne, test/dynamo/test_dicts.py::DictGuardTests::test_popitem, test/dynamo/test_dicts.py::DictMethodsTests::test_binop_or, test/dynamo/test_dicts.py::DictMethodsTests::test_clear, test/dynamo/test_dicts.py::DictMethodsTests::test_cmp_eq, test/dynamo/test_dicts.py::DictMethodsTests::test_cmp_ne, test/dynamo/test_dicts.py::DictMethodsTests::test_copy, test/dynamo/test_dicts.py::DictMethodsTests::test_fromkeys, test/dynamo/test_dicts.py::DictMethodsTests::test_get, test/dynamo/test_dicts.py::DictMethodsTests::test_items, test/dynamo/test_dicts.py::DictMethodsTests::test_keys, test/dynamo/test_dicts.py::DictMethodsTests::test_pop, test/dynamo/test_dicts.py::DictMethodsTests::test_popitem, test/dynamo/test_dicts.py::DictMethodsTests::test_setdefault, test/dynamo/test_dicts.py::DictMethodsTests::test_update, test/dynamo/test_dicts.py::DictMethodsTests::test_values, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_binop_or, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_clear, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_cmp_eq, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_cmp_ne, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_copy, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_fromkeys, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_get, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_items, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_keys, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_pop, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_popitem, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_setdefault, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_update, test/dynamo/test_dicts.py::DictSubclassMethodsTests::test_values, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_binop_or, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_clear, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_cmp_eq, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_cmp_eq_order, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_cmp_ne, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_copy, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_fromkeys, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_get, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_items, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_keys, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_pop, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_popitem, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_setdefault, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_update, test/dynamo/test_dicts.py::OrderedDictMethodsTests::test_values 2025-07-17T09:43:45.0215357Z 2025-07-17T09:43:45.0215497Z Running export/test_serialize 1/1 ... [2025-07-17 09:43:45.018061] 2025-07-17T09:43:45.0215774Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:43:45.0216442Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_serialize.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:43:45.018352] 2025-07-17T09:44:27.7619366Z 2025-07-17T09:44:27.7620376Z export/test_serialize 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_serialize_1.1_b0516840f6da6804_.log 2025-07-17T09:44:27.7640111Z Running 98 items in this shard: test/export/test_serialize.py::TestSerialize::test_canonicalize, test/export/test_serialize.py::TestSerialize::test_export_example_inputs_preserved, test/export/test_serialize.py::TestSerialize::test_export_with_extension_op_serialization, test/export/test_serialize.py::TestSerialize::test_int_list, test/export/test_serialize.py::TestSerialize::test_kwargs_default, test/export/test_serialize.py::TestSerialize::test_metadata_parsing_with_layer_split, test/export/test_serialize.py::TestSerialize::test_metadata_run_decomp_serder, test/export/test_serialize.py::TestSerialize::test_multi_return_some_unused, test/export/test_serialize.py::TestSerialize::test_nested_layer_split, test/export/test_serialize.py::TestSerialize::test_nonfinite_inputs, test/export/test_serialize.py::TestSerialize::test_predispatch_export_with_autograd_op, test/export/test_serialize.py::TestSerialize::test_rational_ranges, test/export/test_serialize.py::TestSerialize::test_serialize_constant_outputs, test/export/test_serialize.py::TestSerialize::test_serialize_infinite_sym_int, test/export/test_serialize.py::TestSerialize::test_serialize_list_returns, test/export/test_serialize.py::TestSerialize::test_serialize_multiple_returns_from_node, test/export/test_serialize.py::TestSerialize::test_serialize_sym_float, test/export/test_serialize.py::TestSerialize::test_serialize_sym_int, test/export/test_serialize.py::TestSerialize::test_symint_list, test/export/test_serialize.py::TestDeserialize::test_arg_from, test/export/test_serialize.py::TestDeserialize::test_auto_functionalize, test/export/test_serialize.py::TestDeserialize::test_basic, test/export/test_serialize.py::TestDeserialize::test_cond, test/export/test_serialize.py::TestDeserialize::test_constraints, test/export/test_serialize.py::TestDeserialize::test_custom_obj, test/export/test_serialize.py::TestDeserialize::test_custom_obj_list_out, test/export/test_serialize.py::TestDeserialize::test_custom_obj_tuple_out, test/export/test_serialize.py::TestDeserialize::test_device, test/export/test_serialize.py::TestDeserialize::test_dynamic, test/export/test_serialize.py::TestDeserialize::test_export_no_inputs, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_assume_constant_result, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_autograd_function, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_class_method, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_cond_branch_class_method, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_cond_branch_nested_function, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_cond_branch_nonlocal_variables, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_cond_closed_over_variable, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_cond_operands, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_cond_predicate, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_constrain_as_size_example, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_constrain_as_value_example, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_decorator, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_dictionary, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_dynamic_shape_assert, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_dynamic_shape_constructor, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_dynamic_shape_if_guard, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_dynamic_shape_map, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_dynamic_shape_slicing, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_dynamic_shape_view, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_fn_with_kwargs, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_list_contains, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_list_unpack, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_nested_function, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_null_context_manager, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_pytree_flatten, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_scalar_output, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_specialized_attribute, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_static_for_loop, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_static_if, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_tensor_setattr, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_type_reflection_method, test/export/test_serialize.py::TestDeserialize::test_exportdb_supported_case_user_input_mutation, test/export/test_serialize.py::TestDeserialize::test_get_attr, test/export/test_serialize.py::TestDeserialize::test_get_attr_list, test/export/test_serialize.py::TestDeserialize::test_hoo_symint_input, test/export/test_serialize.py::TestDeserialize::test_list_of_optional_tensors, test/export/test_serialize.py::TestDeserialize::test_map, test/export/test_serialize.py::TestDeserialize::test_module, test/export/test_serialize.py::TestDeserialize::test_module_meta, test/export/test_serialize.py::TestDeserialize::test_multi_return, test/export/test_serialize.py::TestDeserialize::test_multiple_getitem, test/export/test_serialize.py::TestDeserialize::test_none_input, test/export/test_serialize.py::TestDeserialize::test_optional_tuple, test/export/test_serialize.py::TestDeserialize::test_positional_argument_with_default_value, test/export/test_serialize.py::TestDeserialize::test_pytree_namedtuple, test/export/test_serialize.py::TestDeserialize::test_serialize_float8, test/export/test_serialize.py::TestDeserialize::test_shape, test/export/test_serialize.py::TestDeserialize::test_sym_bool, test/export/test_serialize.py::TestDeserialize::test_sym_bool_dynamic_shapes, test/export/test_serialize.py::TestDeserialize::test_sym_bool_torch_check_equal, test/export/test_serialize.py::TestDeserialize::test_sym_float, test/export/test_serialize.py::TestDeserialize::test_sym_int_torch_check_equal, test/export/test_serialize.py::TestDeserialize::test_sym_ite, test/export/test_serialize.py::TestDeserialize::test_tensor_tensor_list, test/export/test_serialize.py::TestDeserialize::test_unbacked_bindings_serialize, test/export/test_serialize.py::TestSchemaVersioning::test_error, test/export/test_serialize.py::TestSaveLoad::test_save_buffer, test/export/test_serialize.py::TestSaveLoad::test_save_constants, test/export/test_serialize.py::TestSaveLoad::test_save_extra, test/export/test_serialize.py::TestSaveLoad::test_save_file, test/export/test_serialize.py::TestSaveLoad::test_save_path, test/export/test_serialize.py::TestSaveLoad::test_version_error, test/export/test_serialize.py::TestSerializeCustomClass::test_custom_class, test/export/test_serialize.py::TestSerializeCustomClass::test_custom_class_containing_fake_tensor, test/export/test_serialize.py::TestSerializeCustomClass::test_custom_class_input_to_function, test/export/test_serialize.py::TestSerializeCustomClass::test_custom_tag_metadata_copy, test/export/test_serialize.py::TestSerializeCustomClass::test_custom_tag_metadata_decomp, test/export/test_serialize.py::TestSerializeCustomClass::test_custom_tag_metadata_serialization 2025-07-17T09:44:27.7658549Z 2025-07-17T09:44:27.7658680Z Running export/test_sparse 1/1 ... [2025-07-17 09:44:27.761820] 2025-07-17T09:44:27.7658951Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:44:27.7659604Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_sparse.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:44:27.762107] 2025-07-17T09:44:31.1335105Z 2025-07-17T09:44:31.1336122Z export/test_sparse 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_sparse_1.1_67b97f91f02d2493_.log 2025-07-17T09:44:31.1375194Z Running 203 items in this shard: test/export/test_sparse.py::TestSparseProp::test_activation_coo, test/export/test_sparse.py::TestSparseProp::test_activation_csr, test/export/test_sparse.py::TestSparseProp::test_add, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_bfloat16_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_bfloat16_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_bfloat16_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_bfloat16_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_bfloat16_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_bfloat16_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_bfloat16_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_bfloat16_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_bfloat16_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_bfloat16_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float16_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float16_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float16_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float16_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float16_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float16_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float16_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float16_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float16_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float16_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float32_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float32_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float32_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float32_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float32_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float32_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float32_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float32_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float32_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float32_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float64_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float64_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float64_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float64_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float64_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float64_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float64_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float64_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float64_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_float64_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_int64_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_int64_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_int64_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_int64_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_int64_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_int64_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_int64_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_int64_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_int64_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_eltwisenet_int64_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_idnet_bfloat16_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_idnet_bfloat16_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_idnet_bfloat16_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_idnet_bfloat16_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_idnet_bfloat16_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_idnet_bfloat16_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_idnet_bfloat16_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_idnet_bfloat16_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_idnet_bfloat16_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_idnet_bfloat16_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_idnet_float16_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_idnet_float16_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_idnet_float16_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_idnet_float16_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_idnet_float16_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_idnet_float16_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_idnet_float16_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_idnet_float16_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_idnet_float16_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_idnet_float16_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_idnet_float32_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_idnet_float32_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_idnet_float32_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_idnet_float32_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_idnet_float32_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_idnet_float32_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_idnet_float32_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_idnet_float32_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_idnet_float32_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_idnet_float32_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_idnet_float64_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_idnet_float64_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_idnet_float64_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_idnet_float64_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_idnet_float64_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_idnet_float64_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_idnet_float64_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_idnet_float64_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_idnet_float64_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_idnet_float64_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_idnet_int64_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_idnet_int64_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_idnet_int64_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_idnet_int64_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_idnet_int64_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_idnet_int64_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_idnet_int64_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_idnet_int64_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_idnet_int64_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_idnet_int64_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_bfloat16_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_bfloat16_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_bfloat16_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_sumnet_bfloat16_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_bfloat16_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_bfloat16_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_bfloat16_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_bfloat16_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_sumnet_bfloat16_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_bfloat16_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_float16_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_float16_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_float16_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_sumnet_float16_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_float16_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_float16_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_float16_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_float16_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_sumnet_float16_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_float16_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_float32_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_float32_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_float32_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_sumnet_float32_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_float32_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_float32_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_float32_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_float32_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_sumnet_float32_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_float32_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_float64_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_float64_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_float64_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_sumnet_float64_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_float64_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_float64_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_float64_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_float64_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_sumnet_float64_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_float64_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_int64_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_int64_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_int64_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_sumnet_int64_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_int64_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_int64_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_int64_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_sumnet_int64_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_sumnet_int64_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_sumnet_int64_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_bfloat16_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_bfloat16_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_bfloat16_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_todensenet_bfloat16_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_bfloat16_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_bfloat16_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_bfloat16_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_bfloat16_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_todensenet_bfloat16_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_bfloat16_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_float16_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_float16_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_float16_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_todensenet_float16_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_float16_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_float16_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_float16_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_float16_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_todensenet_float16_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_float16_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_float32_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_float32_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_float32_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_todensenet_float32_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_float32_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_float32_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_float32_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_float32_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_todensenet_float32_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_float32_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_float64_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_float64_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_float64_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_todensenet_float64_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_float64_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_float64_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_float64_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_float64_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_todensenet_float64_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_float64_int64_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_int64_int32_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_int64_int32_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_int64_int32_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_todensenet_int64_int32_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_int64_int32_SparseCSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_int64_int64_SparseBSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_int64_int64_SparseBSR, test/export/test_sparse.py::TestSparseProp::test_todensenet_int64_int64_SparseCOO, test/export/test_sparse.py::TestSparseProp::test_todensenet_int64_int64_SparseCSC, test/export/test_sparse.py::TestSparseProp::test_todensenet_int64_int64_SparseCSR 2025-07-17T09:44:31.1413347Z 2025-07-17T09:44:31.1413479Z Running export/test_swap 1/1 ... [2025-07-17 09:44:31.133562] 2025-07-17T09:44:31.1413755Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:44:31.1414423Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_swap.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:44:31.133854] 2025-07-17T09:44:42.1712228Z 2025-07-17T09:44:42.1713854Z export/test_swap 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_swap_1.1_14e13d3f573b2089_.log 2025-07-17T09:44:42.1717475Z Running 18 items in this shard: test/export/test_swap.py::TestSwap_nonstrict::test_custom_input_args, test/export/test_swap.py::TestSwap_nonstrict::test_custom_input_kwargs, test/export/test_swap.py::TestSwap_nonstrict::test_custom_output, test/export/test_swap.py::TestSwap_nonstrict::test_dedup_sym_size, test/export/test_swap.py::TestSwap_nonstrict::test_nested_leaf, test/export/test_swap.py::TestSwap_nonstrict::test_remove_duplicate_pytree_different_order, test/export/test_swap.py::TestSwap_nonstrict::test_remove_duplicate_pytree_simple, test/export/test_swap.py::TestSwap_nonstrict::test_unflatten_preserve_signature, test/export/test_swap.py::TestSwap_nonstrict::test_unflatten_preserve_with_unused_input, test/export/test_swap.py::TestSwap_strict::test_custom_input_args, test/export/test_swap.py::TestSwap_strict::test_custom_input_kwargs, test/export/test_swap.py::TestSwap_strict::test_custom_output, test/export/test_swap.py::TestSwap_strict::test_dedup_sym_size, test/export/test_swap.py::TestSwap_strict::test_nested_leaf, test/export/test_swap.py::TestSwap_strict::test_remove_duplicate_pytree_different_order, test/export/test_swap.py::TestSwap_strict::test_remove_duplicate_pytree_simple, test/export/test_swap.py::TestSwap_strict::test_unflatten_preserve_signature, test/export/test_swap.py::TestSwap_strict::test_unflatten_preserve_with_unused_input 2025-07-17T09:44:42.1720805Z 2025-07-17T09:44:42.1721054Z Running export/test_tools 1/1 ... [2025-07-17 09:44:42.171021] 2025-07-17T09:44:42.1721410Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:44:42.1722088Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_tools.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:44:42.171304] 2025-07-17T09:44:49.8020418Z 2025-07-17T09:44:49.8021550Z export/test_tools 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_tools_1.1_116aa6c0285f1d5a_.log 2025-07-17T09:44:49.8022417Z Running 2 items in this shard: test/export/test_tools.py::TestExportTools::test_report_exportability_basic, test/export/test_tools.py::TestExportTools::test_report_exportability_with_issues 2025-07-17T09:44:49.8022905Z 2025-07-17T09:44:49.8023063Z Running export/test_torchbind 1/1 ... [2025-07-17 09:44:49.801796] 2025-07-17T09:44:49.8023346Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:44:49.8024723Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_torchbind.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:44:49.802149] 2025-07-17T09:45:04.2951790Z 2025-07-17T09:45:04.2952815Z export/test_torchbind 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_torchbind_1.1_2f86fd73dd7ecaba_.log 2025-07-17T09:45:04.2970404Z Running 70 items in this shard: test/export/test_torchbind.py::TestExportTorchbind::test_aot_export_tensor_queue_operators, test/export/test_torchbind.py::TestExportTorchbind::test_attribute_as_custom_op_argument_pre_dispatch_False, test/export/test_torchbind.py::TestExportTorchbind::test_attribute_as_custom_op_argument_pre_dispatch_True, test/export/test_torchbind.py::TestExportTorchbind::test_attribute_pre_dispatch_False, test/export/test_torchbind.py::TestExportTorchbind::test_attribute_pre_dispatch_True, test/export/test_torchbind.py::TestExportTorchbind::test_custom_obj_list_out_pre_dispatch_False, test/export/test_torchbind.py::TestExportTorchbind::test_custom_obj_list_out_pre_dispatch_True, test/export/test_torchbind.py::TestExportTorchbind::test_custom_obj_tuple_out_pre_dispatch_False, test/export/test_torchbind.py::TestExportTorchbind::test_custom_obj_tuple_out_pre_dispatch_True, test/export/test_torchbind.py::TestExportTorchbind::test_custom_obj_unbacked_symint_pre_dispatch_False, test/export/test_torchbind.py::TestExportTorchbind::test_custom_obj_unbacked_symint_pre_dispatch_True, test/export/test_torchbind.py::TestExportTorchbind::test_deepcopy, test/export/test_torchbind.py::TestExportTorchbind::test_export_inplace_custom_op, test/export/test_torchbind.py::TestExportTorchbind::test_identifying_torchbind_ops, test/export/test_torchbind.py::TestExportTorchbind::test_input_as_custom_op_argument_pre_dispatch_False, test/export/test_torchbind.py::TestExportTorchbind::test_input_as_custom_op_argument_pre_dispatch_True, test/export/test_torchbind.py::TestExportTorchbind::test_input_pre_dispatch_False, test/export/test_torchbind.py::TestExportTorchbind::test_input_pre_dispatch_True, test/export/test_torchbind.py::TestExportTorchbind::test_make_fx_schema_checking_script_object, test/export/test_torchbind.py::TestExportTorchbind::test_make_fx_tensor_queue_methods_fakify_internal_states_make_fx_tracing_mode_fake, test/export/test_torchbind.py::TestExportTorchbind::test_make_fx_tensor_queue_methods_fakify_internal_states_make_fx_tracing_mode_symbolic, test/export/test_torchbind.py::TestExportTorchbind::test_make_fx_tensor_queue_methods_make_fx_tracing_mode_fake, test/export/test_torchbind.py::TestExportTorchbind::test_make_fx_tensor_queue_methods_make_fx_tracing_mode_symbolic, test/export/test_torchbind.py::TestExportTorchbind::test_make_fx_tensor_queue_operators_fallthrough_via_lib_impl, test/export/test_torchbind.py::TestExportTorchbind::test_make_fx_tensor_queue_operators_fallthrough_via_py_impl, test/export/test_torchbind.py::TestExportTorchbind::test_method_schema, test/export/test_torchbind.py::TestExportTorchbind::test_non_strict_export_methods, test/export/test_torchbind.py::TestExportTorchbind::test_none_pre_dispatch_False, test/export/test_torchbind.py::TestExportTorchbind::test_none_pre_dispatch_True, test/export/test_torchbind.py::TestExportTorchbind::test_safe_to_trace_with_real, test/export/test_torchbind.py::TestExportTorchbind::test_torchbind_alias_pre_dispatch_False, test/export/test_torchbind.py::TestExportTorchbind::test_torchbind_alias_pre_dispatch_True, test/export/test_torchbind.py::TestExportTorchbind::test_torchbind_input_and_alias_pre_dispatch_False, test/export/test_torchbind.py::TestExportTorchbind::test_torchbind_input_and_alias_pre_dispatch_True, test/export/test_torchbind.py::TestExportTorchbind::test_torchbind_op_fallthrough_keys_respects_lib_impl, test/export/test_torchbind.py::TestExportTorchbind::test_torchbind_op_register_fallthrough, test/export/test_torchbind.py::TestExportTorchbind::test_unlift_custom_obj_pre_dispatch_False, test/export/test_torchbind.py::TestExportTorchbind::test_unlift_custom_obj_pre_dispatch_True, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_body_aliasing_contents_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_body_aliasing_contents_backend_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_error_on_input_aliasing_contents_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_error_on_input_aliasing_contents_backend_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_error_on_non_fakified_method_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_error_on_non_fakified_method_backend_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_error_on_script_obj_missing_attr_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_error_on_script_obj_missing_attr_backend_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_error_on_script_obj_setattr_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_error_on_script_obj_setattr_backend_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_global_obj_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_global_obj_backend_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_obj_as_hop_input_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_obj_as_hop_input_backend_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_obj_attributes_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_obj_attributes_backend_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_obj_closure_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_obj_closure_backend_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_obj_graph_breaks, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_obj_torchbind_op_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_obj_torchbind_op_backend_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_script_object_input_automatic_dynamic_shape, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_script_object_input_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_script_object_input_backend_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_script_object_input_guards_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_script_object_input_guards_backend_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_tensor_op_in_tensor_flatten_backend_aot_eager, test/export/test_torchbind.py::TestCompileTorchbind::test_compile_tensor_op_in_tensor_flatten_backend_eager, test/export/test_torchbind.py::TestRegisterFakeClass::test_register_fake_class_from_real_not_classmethod, test/export/test_torchbind.py::TestRegisterFakeClass::test_register_fake_class_no_from_real, test/export/test_torchbind.py::TestRegisterFakeClass::test_register_fake_class_no_torch_bind_class, test/export/test_torchbind.py::TestRegisterFakeClass::test_register_fake_class_valid 2025-07-17T09:45:04.2986866Z 2025-07-17T09:45:04.2987010Z Running export/test_tree_utils 1/1 ... [2025-07-17 09:45:04.295598] 2025-07-17T09:45:04.2987297Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:45:04.2987980Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_tree_utils.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:45:04.295886] 2025-07-17T09:45:07.9180990Z 2025-07-17T09:45:07.9182650Z export/test_tree_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_tree_utils_1.1_3ec52c1203acd439_.log 2025-07-17T09:45:07.9183480Z Running 2 items in this shard: test/export/test_tree_utils.py::TestTreeUtils::test_equivalence_check, test/export/test_tree_utils.py::TestTreeUtils::test_reorder_kwargs 2025-07-17T09:45:07.9183887Z 2025-07-17T09:45:07.9188571Z Running export/test_unflatten 1/1 ... [2025-07-17 09:45:07.917888] 2025-07-17T09:45:07.9188895Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:45:07.9189565Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_unflatten.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:45:07.918175] 2025-07-17T09:45:27.3164779Z 2025-07-17T09:45:27.3165769Z export/test_unflatten 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_unflatten_1.1_dc00dd6281c068c4_.log 2025-07-17T09:45:27.3189488Z Running 27 items in this shard: test/export/test_unflatten.py::TestUnflatten::test_attr_as_submod_input, test/export/test_unflatten.py::TestUnflatten::test_dedup_sym_size, test/export/test_unflatten.py::TestUnflatten::test_double_nested_submodule, test/export/test_unflatten.py::TestUnflatten::test_duplicate_placeholder, test/export/test_unflatten.py::TestUnflatten::test_fx_trace, test/export/test_unflatten.py::TestUnflatten::test_nested_leaf_non_strict, test/export/test_unflatten.py::TestUnflatten::test_placeholder_and_get_attr_ordering_after_unflattened, test/export/test_unflatten.py::TestUnflatten::test_simple_alias, test/export/test_unflatten.py::TestUnflatten::test_unflatten_buffer_mutation, test/export/test_unflatten.py::TestUnflatten::test_unflatten_constant_obj, test/export/test_unflatten.py::TestUnflatten::test_unflatten_constant_tensor, test/export/test_unflatten.py::TestUnflatten::test_unflatten_container_type, test/export/test_unflatten.py::TestUnflatten::test_unflatten_eager, test/export/test_unflatten.py::TestUnflatten::test_unflatten_empty_branch, test/export/test_unflatten.py::TestUnflatten::test_unflatten_nested, test/export/test_unflatten.py::TestUnflatten::test_unflatten_nested_access, test/export/test_unflatten.py::TestUnflatten::test_unflatten_none, test/export/test_unflatten.py::TestUnflatten::test_unflatten_param_list_dict, test/export/test_unflatten.py::TestUnflatten::test_unflatten_preserve_signature, test/export/test_unflatten.py::TestUnflatten::test_unflatten_preserve_with_unused_input, test/export/test_unflatten.py::TestUnflatten::test_unflatten_requires_grad_param, test/export/test_unflatten.py::TestUnflatten::test_unflatten_shared_submodule, test/export/test_unflatten.py::TestUnflatten::test_unflatten_skipped_call_module, test/export/test_unflatten.py::TestUnflatten::test_unflatten_submodule_ordering, test/export/test_unflatten.py::TestUnflatten::test_unflatten_with_inplace_compile, test/export/test_unflatten.py::TestUnflatten::test_unflatten_wrong_input, test/export/test_unflatten.py::TestUnflatten::test_unflattened_module_nodes_has_meta_val 2025-07-17T09:45:27.3195017Z 2025-07-17T09:45:27.3195190Z Running export/test_unflatten_training_ir 1/1 ... [2025-07-17 09:45:27.316295] 2025-07-17T09:45:27.3195497Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:45:27.3196208Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_unflatten_training_ir.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:45:27.316590] 2025-07-17T09:45:41.9584987Z 2025-07-17T09:45:41.9586209Z export/test_unflatten_training_ir 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_unflatten_training_ir_1.1_bd77d4cc577ca6df_.log 2025-07-17T09:45:41.9595989Z Running 27 items in this shard: test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_attr_as_submod_input_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_dedup_sym_size_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_double_nested_submodule_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_duplicate_placeholder_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_fx_trace_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_nested_leaf_non_strict_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_placeholder_and_get_attr_ordering_after_unflattened_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_simple_alias_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_buffer_mutation_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_constant_obj_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_constant_tensor_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_container_type_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_eager_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_empty_branch_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_nested_access_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_nested_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_none_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_param_list_dict_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_preserve_signature_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_preserve_with_unused_input_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_requires_grad_param_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_shared_submodule_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_skipped_call_module_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_submodule_ordering_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_with_inplace_compile_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflatten_wrong_input_training_ir, test/export/test_unflatten_training_ir.py::TrainingIRUnflattenTestUnflatten::test_unflattened_module_nodes_has_meta_val_training_ir 2025-07-17T09:45:41.9604240Z 2025-07-17T09:45:41.9604376Z Running export/test_upgrader 1/1 ... [2025-07-17 09:45:41.958289] 2025-07-17T09:45:41.9604660Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:45:41.9605342Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_upgrader.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:45:41.958579] 2025-07-17T09:45:45.1797117Z 2025-07-17T09:45:45.1798709Z export/test_upgrader 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_upgrader_1.1_7ac6b63b6fc59715_.log 2025-07-17T09:45:45.1800549Z Running 6 items in this shard: test/export/test_upgrader.py::TestUpgrader::test_field_renaming_chain_from_v0_complete, test/export/test_upgrader.py::TestUpgrader::test_field_renaming_chain_from_v0_missing_field, test/export/test_upgrader.py::TestUpgrader::test_field_renaming_from_v1_partial_chain, test/export/test_upgrader.py::TestUpgrader::test_nn_module_stack_error_handling_invalid_type, test/export/test_upgrader.py::TestUpgrader::test_nn_module_stack_transformation_from_v0, test/export/test_upgrader.py::TestUpgrader::test_nodes_without_metadata_handled_gracefully 2025-07-17T09:45:45.1802488Z 2025-07-17T09:45:45.1802646Z Running export/test_verifier 1/1 ... [2025-07-17 09:45:45.179549] 2025-07-17T09:45:45.1802924Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:45:45.1803588Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'export/test_verifier.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:45:45.179839] 2025-07-17T09:45:54.8140275Z 2025-07-17T09:45:54.8141866Z export/test_verifier 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_verifier_1.1_4cc5e0b759c02688_.log 2025-07-17T09:45:54.8144746Z Running 10 items in this shard: test/export/test_verifier.py::TestVerifier::test_ep_verifier_basic, test/export/test_verifier.py::TestVerifier::test_ep_verifier_buffer_mutate, test/export/test_verifier.py::TestVerifier::test_ep_verifier_invalid_buffer, test/export/test_verifier.py::TestVerifier::test_ep_verifier_invalid_output, test/export/test_verifier.py::TestVerifier::test_ep_verifier_invalid_param, test/export/test_verifier.py::TestVerifier::test_verifier_basic, test/export/test_verifier.py::TestVerifier::test_verifier_call_module, test/export/test_verifier.py::TestVerifier::test_verifier_higher_order, test/export/test_verifier.py::TestVerifier::test_verifier_nested_invalid_module, test/export/test_verifier.py::TestVerifier::test_verifier_no_functional 2025-07-17T09:45:54.8146570Z 2025-07-17T09:45:54.8146722Z Running functorch/test_ac_knapsack 1/1 ... [2025-07-17 09:45:54.813795] 2025-07-17T09:45:54.8147036Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:45:54.8147726Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'functorch/test_ac_knapsack.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:45:54.814082] 2025-07-17T09:45:58.4856074Z 2025-07-17T09:45:58.4857059Z functorch/test_ac_knapsack 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ac_knapsack_1.1_fd550ecca72448bb_.log 2025-07-17T09:45:58.4861867Z Running 15 items in this shard: test/functorch/test_ac_knapsack.py::TestGraphInfoProvider::test_full_joint_nx_graph, test/functorch/test_ac_knapsack.py::TestGraphInfoProvider::test_get_knapsack_memory_input, test/functorch/test_ac_knapsack.py::TestGraphInfoProvider::test_get_knapsack_runtime_input, test/functorch/test_ac_knapsack.py::TestGraphInfoProvider::test_get_non_ac_peak_memory, test/functorch/test_ac_knapsack.py::TestGraphInfoProvider::test_get_theoretical_max_runtime, test/functorch/test_ac_knapsack.py::TestGraphInfoProvider::test_inialize_from_graph, test/functorch/test_ac_knapsack.py::TestGraphInfoProvider::test_recomputable_node_only_graph, test/functorch/test_ac_knapsack.py::TestGraphInfoProvider::test_recomputable_node_only_graph_with_larger_graph_context, test/functorch/test_ac_knapsack.py::TestGraphInfoProvider::test_simplified_fx_joint_graph, test/functorch/test_ac_knapsack.py::TestKnapsackEvaluator::test_evaluate_distribution_of_results_for_knapsack_algo, test/functorch/test_ac_knapsack.py::TestKnapsackEvaluator::test_evaluate_knapsack_output_accounting_for_backward_pass, test/functorch/test_ac_knapsack.py::TestKnapsackEvaluator::test_evaluate_knapsack_output_not_accounting_for_backward_pass, test/functorch/test_ac_knapsack.py::TestKnapsackEvaluator::test_evaluate_knapsack_output_with_wrong_sized_values, test/functorch/test_ac_knapsack.py::TestKnapsackEvaluator::test_get_backward_memory_from_topologically_sorted_graph, test/functorch/test_ac_knapsack.py::TestKnapsackEvaluator::test_get_knee_point_memory_budget 2025-07-17T09:45:58.4865942Z 2025-07-17T09:45:58.4866093Z Running functorch/test_ac_logging 1/1 ... [2025-07-17 09:45:58.485434] 2025-07-17T09:45:58.4866384Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:45:58.4867073Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'functorch/test_ac_logging.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:45:58.485729] 2025-07-17T09:46:01.8569648Z 2025-07-17T09:46:01.8570764Z functorch/test_ac_logging 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ac_logging_1.1_4d6ce8782b78d83b_.log 2025-07-17T09:46:01.8572206Z Running 4 items in this shard: test/functorch/test_ac_logging.py::TestAcLogging::test_create_activation_checkpointing_logging_structure_payload, test/functorch/test_ac_logging.py::TestAcLogging::test_create_joint_graph_edges, test/functorch/test_ac_logging.py::TestAcLogging::test_create_joint_graph_node_information, test/functorch/test_ac_logging.py::TestAcLogging::test_create_structured_trace_for_min_cut_info 2025-07-17T09:46:01.8573872Z 2025-07-17T09:46:01.8574037Z Running functorch/test_control_flow 1/1 ... [2025-07-17 09:46:01.856771] 2025-07-17T09:46:01.8574330Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:46:01.8575007Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'functorch/test_control_flow.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:46:01.857061] 2025-07-17T09:50:10.8814258Z 2025-07-17T09:50:10.8815257Z dynamo/test_dynamic_shapes 1/3 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_dynamic_shapes_1.3_c0209c27f57f4cb1_.log 2025-07-17T09:50:10.8998382Z Running 623 items in this shard: test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_autocast_arguments_binding_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_autocast_cpu_graph_break_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_autocast_decorator_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_autocast_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_autograd_profiler_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_context_wrapping_grad_mode_decorator_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_context_wrapping_grad_mode_nested_function_decorator_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_cuda_amp_autocast_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_cuda_event_reconstruct_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_cuda_stream_compared_with_constant_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_cuda_stream_compared_with_stream_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_cuda_stream_context_manager2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_disable_saved_tensors_hooks_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_disable_saved_tensors_hooks_prev_disabled_nested_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_generic_context_manager_CustomizedCtxManager_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_generic_context_manager_customized_ctx_manager_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_generic_context_manager_with_graph_break_customized_ctx_manager_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_grad_mode_guard_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_inactive_context_graph_break_local_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_inactive_context_graph_break_local_nullctx_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_inactive_context_graph_break_stack_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_nested_generic_context_manager_CustomizedCtxManager_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_nested_grad_mode_graph_break_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_no_grad_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_sdpa_kernel_ctx_manager1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesCtxManagerTests::test_sdpa_kernel_ctx_manager3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_T_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_addcdiv__dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_addcdiv_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_call_dict1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_callable_builtin_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_cls_hasattr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_compare_constant_and_tensor_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_constant2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_constant3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_constant_set_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_cublas_allow_tf32_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_default_dict_lambda_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_default_dict_list_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_defaultdict_setdefault3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_dict_hasattr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_dict_id_guard_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_dict_items_sorted_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_dict_keys_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_dict_mutable_map_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_dict_setdefault1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_dict_setdefault3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_distributed_is_initialized_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_dtype_compare_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_filter_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_filter_graph_break_reconstruct_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_filter_infinite_iterator_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_fstrings1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_fstrings2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_functools_cache_guard_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_functools_partial_binding_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_generic_namedtuple_hasattr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_generic_namedtuple_subclass_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_generic_namedtuple_user_methods_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_get_autocast_gpu_dtype_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_get_default_dtype_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_get_device_properties_tensor_device_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_globalfn_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_indexed_range_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_inline_jit__unwrap_optional_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_inline_with_default_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_inner_function_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_is_floating_point_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_is_not_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_is_not_null_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_islice_chain_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_itertools_chain_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_itertools_combinations_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_jit_annotate_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_len_constant_list_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_list_add_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_list_compare_polyfill_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_list_reversed_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_list_slice_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_list_truth_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_listarg1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_mT_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_manual_seed_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_map_call_function_ex_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_map_infinite_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_map_reconstruct_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_map_return_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_map_tuple_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_map_unpack_twice_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_map_unpack_vars_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_map_zip_dict_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_math_radians_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_mean_sum_np_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_methodcall2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_methodcall3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_module_constant_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_namedtuple_hasattr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_namedtuple_subclass_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_namedtuple_user_methods_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_ndim_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_no_recompile_inner_lambda_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_non_inlined_closure_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_not_list_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_np_constant_collections_guards_float_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_np_iinfo_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_number_method_method_as_integer_ratio_num_type3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_number_method_method_bit_length_num_type1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_number_method_method_conjugate_num_type4_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_numpy_linalg_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_numpy_random_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_as_input_partials_lambda_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_graph_break_reconstruct_args_and_kwargs_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_hasattr_attr___annotations___dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_hasattr_attr___class___dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_hasattr_attr___delattr___dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_hasattr_attr___dir___dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_hasattr_attr___doc___dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_hasattr_attr___getattribute___dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_hasattr_attr___kwdefaults___dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_hasattr_attr___ne___dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_hasattr_attr___reduce___dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_hasattr_attr___setattr___dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_hasattr_attr___sizeof___dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_udf_arg_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_udf_kwarg_method_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_partials_udf_kwarg_module_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_pop_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_rand_tensor_partial_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_range_length_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_range_with_index_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_reduce_with_single_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_return_multiple_numpy_ndarray_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_return_tuple1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_set_keys_view_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_shape1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_size_tuple_add_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_slice2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_slice3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_sliced_range_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_sum_with_start_arg_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_tensor_dim_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_tensor_element_size_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_tensor_is_complex_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_tensor_len_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_tensor_type2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_to_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_truth_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_tuple_iadd_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_two_point_iter_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_unary_fold_op_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_unpack3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_unsqueeze_inplace_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFunctionTests::test_viatorch_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_312_binary_slice_with_graph_break2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_RAISE_VARARGS_0_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_add_to_set_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_anomaly_aot_autograd_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_aot_autograd_propagate_unbacked_symints_shape_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_argwhere_with_dynamic_shapes_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_builtin_bool_on_symbool_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_callpacked_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_catch_watchings1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_cell_captured_by_existing_func_but_not_root_frame_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_cell_output1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_class_duner_mro_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_clone_sparse_input_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_closure_out_of_scope_cell_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_closure_with_mutation_and_graph_break_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_compilation_metrics_size_limit_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_cond_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_cond_export_single_arg_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_cond_nested_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_conditional_list_comp_in_context_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_constant_getattr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_cross_entropy_loss_fancy_ctor1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_cross_entropy_loss_fancy_ctor2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_cross_entropy_loss_simple_ctor_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_custom_module_free_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_dataclass_fields_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_deque_input_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_deterministic_algorithms_mutated_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_dunder_methods_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_dunder_new_function_inlining2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_dunder_weakref_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_dynamic_one_hot_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_dynamic_sources_force_parameter_static_shapes_and_property_static_shapes_override_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_dynamic_sources_graph_break_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_dynamic_sources_int_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_dynamic_sources_tensor_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_dynamo_disabled_in_custom_op_kernels_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_enum_as_dict_key_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_enum_subclass_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_error_on_nested_fx_trace_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_error_on_recompile_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_escaping_closure_var_with_nonlocal_var_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_fn_hasattr__name__3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_frozen_dataclass_default_factory_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_frozen_dataclass_kw_only_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_frozenset_of_non_literals_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_function_annotation_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_generate_tensor_from_list_of_numpy_primitive_type_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_get_attr_function_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_get_custom_tensor_attribute_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_get_instruction_source_311_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_getattr_dict_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_getattrvariable_as_python_constant_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_global_state_guard_serialization_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_graph_break_compilation_metrics_on_failure_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_guard_failure_fn_shape_control_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_guard_filter_fn_by_name_and_value_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_guard_filter_nn_modules_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_guards_cse_pass_single_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_hasattr_nn_module_guard_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_id_guarded_module_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_if_cond_nn_mod3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_inline_closure_returned_by_another_function_and_captures_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_inline_dict_function_passed_as_arg_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_inline_func_jump_on_tensor_condition_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_inline_list_mutation_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_inline_local_dict_clear_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_input_cell_mutation_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_int_int_comparisons_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_int_shape_binops_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_int_shape_comparisons_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_int_shape_inplace_binops_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_intermediary_tensor_grad_access_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_invalid_args_builtin_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_is_floating_point2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_is_tensor2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_is_tensor_like2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_item_changes_new_shape_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_itertools_accumulate_symint_default_sum_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_itertools_accumulate_tensors_default_sum_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_itertools_accumulate_tensors_kwargs_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_itertools_groupby_pure_python_key_func_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_itertools_infinite_repeat_mutation_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_list_hasattr1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_map_side_effects_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_map_with_quantization_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_mark_static_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_mark_unbacked_strict_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_matmul1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_min_max_over_iterable_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_module_complex_iter_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_multiple_inheritance_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_namedtuple3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_namedtuple_with_custom_getitem_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_ne_operator_with_custom_eq_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_ne_operator_with_custom_ne_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_nested_optimize_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_nested_optimize_run_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_nested_sequential_try_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_nn_module_getattr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_nn_module_getattribute_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_nn_sequential_invocation_reposition_indices_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_no_guard_for_unused_sym_node_fstring_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_numpy_gt_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_numpy_int_constant_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_numpy_ndarray_works_with_builtin_function_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_numpy_non_torch_dtype_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_numpy_random_config_to_numpy_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_numpy_ufunc_out_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_numpy_ufunc_out_graph_break_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_object_classmethod_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_out_variants_with_resizing_on_graph_inputs_with_dynamic1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_overridden_getattribute_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_packaging_version_parse_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_param_shape_binops_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_parameter_free_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_precompile_entries_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_precompile_entry_miss_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_pure_python_accumulate_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_python_slice_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_pytree_tree_flatten_unflatten_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_pytree_tree_map_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_raise_guard_indirect_full_constraint_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_raise_guard_partial_constraint_no_graph_break_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_raises_importerror1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_range_iter_side_effects_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_range_with_shape_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_recompile_on_global_state_change_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_reconstruct_frozen_dataclass_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_reconstruct_set_across_graph_break_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_recursion_depth_guards_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_recursive_inline_list_mutation_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_release_module_memory_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_repro_graph_breaks_in__get_item_by_idx_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_running_nested_func_with_captured_tensor_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_sample_input_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_scalar_tensor_is_equivalent_to_symint_argument_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_set_aliasing_recompiles_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_set_discard_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_setattr_mutation1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_setattr_mutation2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_shape_env_equal_constructor_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_shape_env_equal_evaluate_expr_refinement_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_shape_type_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_shape_unpack_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_side_effects_codegen_update_mutated_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_slice_input_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_source_non_input_grad_access_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_storage_return_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_str_format_assert1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_str_format_return2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_structseq2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_sym_constrain_range_on_replaced_unbacked_symbol_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_sym_max_unbacked_sizelike_simplification_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_symint_as_device_kwarg_multi_gpu_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tagging_tensors_simple_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tensor_data_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tensor_dict1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tensor_interacts_with_numpy_ndarray_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tensor_is_contiguous_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tensor_item_capture_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tensor_item_no_capture_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tolist_1d_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tolist_kd_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tolist_scalar_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_top_package_import_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_torch_check_is_size_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_torch_distributions_lazy_property_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_torch_dtype_python_type_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_torch_generator_set_state_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_torch_guards_stack_frame_register_inlining_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_torch_size_numel_dynamic_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_torch_variable_hasattr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_trace_ndarray_frame_2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tracing_nested_py_tree_dicts_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tracing_tree_map_only_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_tuple_iadd_with_shape_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_typing_union_and_optional_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_unbacked_2d_expand_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_unbacked_repeat_cat_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_unbacked_strict_mode_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_unhandled_exception_in_dynamo2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_unhandled_exception_in_dynamo_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_unique_consecutive_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_unpack5_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_update_locals_and_stack_uses_shared_cache_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_user_code_statically_known_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_user_defined_setattr2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_user_getattr1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_usr_cls_staticmethod_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_variable_access_in_exception_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_write_to_closures_in_inlining_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_yield_from_in_a_loop_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_yield_from_user_stop_iteration_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_yield_gen_and_from_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesMiscTests::test_yield_send_to_subgenerator_graph_break_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_abc_setattr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_add_complex_conj_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_amp_foreach_fake_impl_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_ao_fake_quantize_tracing_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_aot_autograd_runtime_wrapper_prologue_profiled_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_as_strided_on_base_with_mutation_works_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_autograd_function_graph_break_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_avoid_dupe_specialization_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_batchnorm_e2e_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_boxes_len_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_chunk_reformer_ff_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_compilation_metrics_on_error_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_convert_boxes_to_pooler_format_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_copy_weird_strides_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_dalle2_maybe_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_deferred_runtime_asserts_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_delattr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_delete_local_error_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_delsubscr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_delsubscr_raises_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_detectron2_instances_cat_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_dynamic_shape_disable_duck_size_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_dynamic_shapes_double_not_equal_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_ellipsis_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_empty_graph_nested_calls_fullgraph_True_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_empty_list_contains_with_jump_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_empty_out_dynamic_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_enum_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_error_return_without_exception_set_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_fsdp_set_input_mutation_applied_when_input_gets_no_gradients_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_grad_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_graph_break_on_jit_isinstance_pep585_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_guard_default_device_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_guard_fail_tensor_bool_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_inductor_dynamic_shapes_broadcasting_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_isinstance_dtype_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_isinstance_storage_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_issue134451_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_list_index_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_list_reverse_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_list_self_reference_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_listcomp_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_maml_item_capture_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_many_overlapping_inputs_does_not_explode_guards_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_many_views_with_mutation_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_maybe_multiply_symint_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_merge_criteria_processor_list2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_nanmean_out_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_nn_module_stack_bc_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_nn_parameter_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_no_tracing_into_eval_frame_ctx_manager_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_nonconst_issubclass_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_nullcontext1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_numpy_not_ndarray_recompiles_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_omegaconf_dictconfig_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_omegaconf_listconfig_iter_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_ones_out_dynamic_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_optimized_module_training_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_out_nested_cell_tuple_shape_change_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_out_none_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_output_aliases_intermediate_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_overwriting_params_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_partially_initialized_module_property_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_partitioner_activation_memory_budget_with_unbacked_symints_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_primtorch_no_graph_break_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_reinplacing_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_requires_grad_guards_with_grad_mode2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_restricted_list_subclass2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_return_value_duplication_mixed_grad_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_return_value_duplication_scalar_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_rewrite_assert_with_msg_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_rewrite_assert_with_non_string_msg_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_rewrite_assert_without_msg_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_setitem_tuple_boolean_mask_diff_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_sigmoid_out2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_size_typematch_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_slice_into_list_mutable_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_slicing_dynamic_shape_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_specialized_stride_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_staticmethod_allow_in_graph_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_string_format_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_super_classmethod_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_super_classmethod_inheritance_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_symint_bitwise_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_tensor_isinstance_tuple_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_tensor_item_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_tensor_random_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_tensor_set_data_backend_aot_eager_func_name_func1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_tensor_set_data_backend_inductor_func_name_func2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_tensor_set_data_mismatched_dtype_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_tensor_split_within_device_cm_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_tensor_uniform_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_trace_functional_tensor_with_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_unbacked_arange_in_bounds_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_user_ctor_ctx_manager_custom_init_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_user_ctor_ctx_manager_custom_init_graph_break_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_user_ctor_ctx_manager_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_user_defined_iter_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_validate_model_kwargs_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_view_dtype_overload_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_weakref_del_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesReproTests::test_zeros_out_dynamic_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_call_fn_with_non_const_inputs_safe_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_cfgmod_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_conv_call_forward_directly_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_fnmembercmp2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_hasattr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_intarg_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_iseval1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_isnonelayer_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_lazy_module5_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_lazy_module6_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_module_attribute_precedence_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_module_class_method_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_module_forward_has_graph_break_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_module_guard_name_is_valid_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_module_property_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_module_static_method_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_moduledict_custom_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_parameterdict_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_parameters2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_sequential_with_duplicated_module2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_super1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_tensorlist_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesNNModuleTests::test_viamodulecall_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_cond_raise_user_error_on_branch_return_multiple_tensors_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_cond_raise_user_error_on_mismatch_return_length_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_cond_raise_user_error_on_mismatch_return_tensor_meta_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_cond_raise_user_error_on_non_tensor_operands_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_cond_raise_user_error_on_unsupported_pred_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_dict_return_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_dict_return_with_aten_graph_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_dupes_2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_dupes_and_bypass_reorder_with_non_tensor_arg_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_dupes_and_bypass_with_non_tensor_output_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_dupes_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_dynamic_slicing_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_dynamic_slicing_simple_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_enforce_equalities_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_decomp_asserts_bad_args_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_dynamic_dim_not_1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_graph_with_complex_reorder_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_mismatched_out_with_aten_graph_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_no_tensor_computation_with_aten_graph_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_raise_guard_partial_constraint_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_shape_control_flow_1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_specialized_int_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_symbolic_shape_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_args_with_default_None_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_args_with_default_float_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_cond_closure_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_constant_free_function_and_class_method_multiarg_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_constant_in_unspecialized_nn_module_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_constant_method_on_module_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_constant_none_control_flow_free_func_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_constant_not_none_control_flow_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_functools_wrapped_method_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_kwargs_and_empty_args_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_kwargs_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_kwargs_with_default_tuple_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_map_zero_sized_tensor_suppress_errors_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_symbool_inputs_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_export_with_wrapped_fn_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_exported_graph_serialization_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_func_return_with_aten_graph_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_invalid_input_nonlocal_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_list_contains_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_list_unpack_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_map_cond_param_buffer_lifted_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_no_tensor_computation_2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_no_tensor_computation_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_param_buffer_safe_from_mutation_recurse_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_param_buffer_safe_from_mutation_simple_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_predispatch_with_higher_order_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_preserve_fx_node_metadata_recompile_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_remove_redundant_dynamic_dim_in_error_message_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_retracibility_dict_container_inp_out_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_round_dynamic_shapes_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_strict_fake_tensor_prop_real_tensors_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_sym_contains_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_symbolic_tracing_within_fake_mode_with_constraints_with_parameters_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_torch_inference_mode_ctx_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_untracked_inputs_in_constraints_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_zeroes_in_and_out_different_shape_on_test_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_zeroes_in_new_shape_scalar_out_permute_dupe_and_bypass_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesExportTests::test_zeroes_in_new_shape_scalar_out_permute_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_capi_call1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_capi_call3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_control_flow1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_dynamic_duck_size_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_dynamic_getitem_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_dynamic_kwarg_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_graph_break_on_item_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_indirect_unsupported3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_resume2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_resume_freevars_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_resume_tuple_iterator_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_resume_with_no_grad3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_stack_state1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_stack_state2_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesSubGraphTests::test_start4_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_capture_global_num_adds_guard_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_capture_untracked_global_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_capture_untracked_global_nested_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_capture_value_created_in_subgraph_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_cond_branches_no_arguments_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_cond_branches_no_arguments_no_closure_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_cond_free_variable_in_both_branches_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_cond_graph_break_in_one_branch_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_cond_pytree_operands_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_cond_pytree_operands_with_non_tensor_leaves_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_fallback_on_graph_break_simple_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_fallback_on_python_primitives_output_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_fn_with_kwargs_in_torch_ops_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_freevars_as_inputs_to_wrap_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_hints_wrapper_incorrect_type_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_hints_wrapper_pytree_inputs_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_lift_tensors_with_shared_symbols_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_map_pytree_return_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_nested_wrap_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_no_freevars_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_return_captured_var_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_return_captured_vars_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_side_effect_del_existing_attr_nonlocal_module_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_side_effect_del_existing_attr_nonlocal_obj_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_side_effect_mutate_global_list_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_side_effect_mutate_global_num_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_side_effect_mutate_nonlocal_num_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_side_effect_mutate_nonlocal_tensor_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_side_effect_set_existing_attr_global_obj_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_side_effect_set_existing_attr_nonlocal_module_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_side_effect_set_new_attr_global_obj_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_side_effect_set_new_attr_nonlocal_obj_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_tensor_with_unbacked_shape_closure_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_unbacked_symbol_closure_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_vmap_source_fn_stack_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_wrap_allow_local_assign_in_body_fn_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_wrap_kwarg_default_else_branch_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_wrap_kwarg_int_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_wrap_pytree_args_not_const_symint_tensor_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesHigherOrderOpTests::test_wrap_source_fn_stack_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_functional_call_sequential_params_and_buffers_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_grad_call_torch_compile_fn_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_grad_has_aux_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_grad_non_tensor_input_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_grad_over_grad_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_jacrev_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_jvp_freevar_python_scalar_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_jvp_freevar_tensor_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_jvp_has_aux_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_linearize_jvp_fn_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_vjp_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_vmap_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_vmap_multiple_outputs_out_dims_tuple_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_vmap_new_tensor_in_body_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_vmap_new_tensor_unused_in_body_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_vmap_side_effects_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesFuncTorchHigherOrderOpTests::test_vmap_with_graph_break_lambda_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_LSTM_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_alias_inputs_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_aot_grad_mode_mutation_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_aot_sequence_nr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_call_fn_with_non_const_inputs_aot_safe_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_different_inputs_overlapping_set_with_mutation_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_donated_buffer4_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_donated_buffer_with_retain_or_create_graph1_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_donated_buffer_with_retain_or_create_graph3_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_eager_sequence_nr_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_negative_testing_mutation_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_no_storage_overlap_guards_no_mutation_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_requires_grad_fake_via_dynamo_recompiles_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesAotAutogradFallbackTests::test_split_with_sizes_aot_autograd_cleans_up_traceback_meta_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesTestSDPA::test_intermediate_attr_access_SDPAParams_dynamic_shapes, test/dynamo/test_dynamic_shapes.py::DynamicShapesTestSDPA::test_returns_SDPAParams_dynamic_shapes 2025-07-17T09:50:10.9161171Z 2025-07-17T09:50:10.9161380Z Running inductor/test_aot_inductor_custom_ops 1/1 ... [2025-07-17 09:50:10.882204] 2025-07-17T09:50:10.9161708Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:50:10.9162417Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_aot_inductor_custom_ops.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:50:10.882503] 2025-07-17T09:52:39.3482273Z 2025-07-17T09:52:39.3483293Z inductor/test_aot_inductor_custom_ops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_custom_ops_1.1_722af083d7acf78a_.log 2025-07-17T09:52:39.3494573Z Running 35 items in this shard: test/inductor/test_aot_inductor_custom_ops.py::AOTInductorLoggingTest::test_shape_env_reuse, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_boxed_run_inputs_clearing_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_add_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_add_output_path_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_all_inputs_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_missing_arg_with_default_value_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_out_variant_without_return_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_return_list_of_single_tensor_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_return_single_tensor_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_square_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_with_concat_inputs_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_with_multiple_outputs_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_custom_op_with_reinterpret_view_inputs_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_fn_with_int_output_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_fn_with_optional_tensor_nullopt_output_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_fn_with_optional_tensor_output_2_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_fn_with_optional_tensor_output_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCpu::test_incorrect_custom_op_schema_cpu, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_boxed_run_inputs_clearing_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_custom_op_add_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_custom_op_add_output_path_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_custom_op_all_inputs_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_custom_op_missing_arg_with_default_value_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_custom_op_out_variant_without_return_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_custom_op_return_list_of_single_tensor_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_custom_op_return_single_tensor_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_custom_op_square_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_custom_op_with_concat_inputs_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_custom_op_with_multiple_outputs_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_custom_op_with_reinterpret_view_inputs_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_fn_with_int_output_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_fn_with_optional_tensor_nullopt_output_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_fn_with_optional_tensor_output_2_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_fn_with_optional_tensor_output_cuda, test/inductor/test_aot_inductor_custom_ops.py::AOTInductorTestABICompatibleCuda::test_incorrect_custom_op_schema_cuda 2025-07-17T09:52:39.3504826Z 2025-07-17T09:52:39.3504991Z GITHUB_RUN_ID, GITHUB_RUN_ATTEMPT, or ARTIFACTS_FILE_SUFFIX not set, not uploading 2025-07-17T09:52:39.3505290Z Uploading artifacts took 0.00 seconds 2025-07-17T09:52:39.3505653Z Running inductor/test_aot_inductor_package 1/1 ... [2025-07-17 09:52:39.348043] 2025-07-17T09:52:39.3505966Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:52:39.3506659Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_aot_inductor_package.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:52:39.348324] 2025-07-17T09:54:17.1979539Z 2025-07-17T09:54:17.1983521Z functorch/test_control_flow 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_control_flow_1.1_5e3fcda76ea1676e_.log 2025-07-17T09:54:17.2435615Z Running 1315 items in this shard: test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_complex, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_different_pytree_output, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_gpu, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_grad_through_cond, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_inner_fn, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_inner_tensor, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_mixed_require_grad, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_nested, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_pytree_input, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_pytree_not_all_inputs_used, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_same_pytree_output, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_simple, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_torch_nn_module, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_user_nn_module, test/functorch/test_control_flow.py::TestControlFlow::test_cond_autograd_zeros_unused_branch_complex_compile_fail_compile_mode_compile_dynamic_shape_scalar_False, test/functorch/test_control_flow.py::TestControlFlow::test_cond_gpu, test/functorch/test_control_flow.py::TestControlFlow::test_cond_in_forloop, test/functorch/test_control_flow.py::TestControlFlow::test_cond_no_trace, test/functorch/test_control_flow.py::TestControlFlow::test_map_autograd_higher_order, test/functorch/test_control_flow.py::TestControlFlow::test_map_autograd_nested_list, test/functorch/test_control_flow.py::TestControlFlow::test_map_autograd_no_grad_output, test/functorch/test_control_flow.py::TestControlFlow::test_map_autograd_simple, test/functorch/test_control_flow.py::TestControlFlow::test_map_autograd_simple_partial_grad, test/functorch/test_control_flow.py::TestControlFlow::test_map_dict_in_out, test/functorch/test_control_flow.py::TestControlFlow::test_map_gpu, test/functorch/test_control_flow.py::TestControlFlow::test_map_illegal_inputs, test/functorch/test_control_flow.py::TestControlFlow::test_map_illegal_outputs, test/functorch/test_control_flow.py::TestControlFlow::test_map_list_in_out, test/functorch/test_control_flow.py::TestControlFlow::test_scan_associative_scan, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_binary_operator_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_carry_carry_alias, test/functorch/test_control_flow.py::TestControlFlow::test_scan_carry_output_alias, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_compile_mode_eager_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_compile_mode_eager_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_compile_mode_none_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_compile_mode_none_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_eager_partial_grad_additional_inputs_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_eager_partial_grad_additional_inputs_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_eager_partial_grad_complex_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_eager_partial_grad_complex_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_eager_partial_grad_init_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_eager_partial_grad_init_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_eager_partial_grad_random_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_eager_partial_grad_random_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_eager_partial_grad_xs_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_eager_partial_grad_xs_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_none_partial_grad_additional_inputs_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_none_partial_grad_additional_inputs_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_none_partial_grad_complex_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_none_partial_grad_complex_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_none_partial_grad_init_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_none_partial_grad_init_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_none_partial_grad_random_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_none_partial_grad_random_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_none_partial_grad_xs_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_False_compile_mode_none_partial_grad_xs_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_eager_partial_grad_additional_inputs_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_eager_partial_grad_additional_inputs_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_eager_partial_grad_complex_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_eager_partial_grad_complex_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_eager_partial_grad_init_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_eager_partial_grad_init_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_eager_partial_grad_random_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_eager_partial_grad_random_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_eager_partial_grad_xs_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_eager_partial_grad_xs_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_none_partial_grad_additional_inputs_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_none_partial_grad_additional_inputs_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_none_partial_grad_complex_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_none_partial_grad_complex_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_none_partial_grad_init_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_none_partial_grad_init_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_none_partial_grad_random_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_none_partial_grad_random_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_none_partial_grad_xs_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_RNN_partial_autograd_reverse_True_compile_mode_none_partial_grad_xs_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_carries_ys_same_grad_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_all_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_additional_inputs_partial_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_for_out_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_equal_grad_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_combine_fn_with_no_grad_init_carries_unequal_grad_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_closure_nested_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_cnt_reverse_False_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_cnt_reverse_False_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_cnt_reverse_True_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_cnt_reverse_True_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_compile_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_complex_pytree_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dim_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_eager_reverse_False_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_eager_reverse_False_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_eager_reverse_False_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_eager_reverse_False_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_eager_reverse_True_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_eager_reverse_True_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_eager_reverse_True_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_eager_reverse_True_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_none_reverse_False_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_none_reverse_False_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_none_reverse_False_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_none_reverse_False_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_none_reverse_True_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_none_reverse_True_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_none_reverse_True_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_matmul_compile_mode_none_reverse_True_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_eager_reverse_False_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_eager_reverse_False_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_eager_reverse_False_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_eager_reverse_False_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_eager_reverse_True_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_eager_reverse_True_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_eager_reverse_True_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_eager_reverse_True_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_none_reverse_False_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_none_reverse_False_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_none_reverse_False_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_none_reverse_False_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_none_reverse_True_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_none_reverse_True_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_none_reverse_True_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_downstream_scan_scan_dim_compile_mode_none_reverse_True_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_eager_cpu_complex64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_eager_cpu_float16, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_eager_cpu_float32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_eager_cpu_int32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_eager_cpu_int64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_eager_cuda_complex64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_eager_cuda_float16, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_eager_cuda_float32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_eager_cuda_int32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_eager_cuda_int64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_none_cpu_complex64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_none_cpu_float16, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_none_cpu_float32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_none_cpu_int32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_none_cpu_int64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_none_cuda_complex64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_none_cuda_float16, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_none_cuda_float32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_none_cuda_int32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_False_compile_mode_none_cuda_int64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_eager_cpu_complex64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_eager_cpu_float16, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_eager_cpu_float32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_eager_cpu_int32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_eager_cpu_int64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_eager_cuda_complex64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_eager_cuda_float16, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_eager_cuda_float32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_eager_cuda_int32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_eager_cuda_int64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_none_cpu_complex64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_none_cpu_float16, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_none_cpu_float32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_none_cpu_int32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_none_cpu_int64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_none_cuda_complex64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_none_cuda_float16, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_none_cuda_float32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_none_cuda_int32, test/functorch/test_control_flow.py::TestControlFlow::test_scan_dtype_reverse_True_compile_mode_none_cuda_int64, test/functorch/test_control_flow.py::TestControlFlow::test_scan_float_output, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_non_tensor, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_pytree_complex_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_scanned_0, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_wrong_pytree_carry_shape, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_wrong_pytree_complex_reverse_False_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_wrong_pytree_complex_reverse_False_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_wrong_pytree_complex_reverse_True_cpu, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_wrong_pytree_complex_reverse_True_cuda, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_wrong_pytree_init_longer_carry, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_wrong_pytree_init_shorter_carry, test/functorch/test_control_flow.py::TestControlFlow::test_scan_init_wrong_shape, test/functorch/test_control_flow.py::TestControlFlow::test_scan_input_carry_alias, test/functorch/test_control_flow.py::TestControlFlow::test_scan_input_mutation, test/functorch/test_control_flow.py::TestControlFlow::test_scan_input_output_alias, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_non_pointwise_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_one_return, test/functorch/test_control_flow.py::TestControlFlow::test_scan_pytree_output, test/functorch/test_control_flow.py::TestControlFlow::test_scan_simple_graph, test/functorch/test_control_flow.py::TestControlFlow::test_scan_simple_graph_wrong_dtype, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_False_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_False_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_False_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_False_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_False_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_False_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_False_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_False_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_True_compile_mode_eager_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_True_compile_mode_eager_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_True_compile_mode_eager_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_True_compile_mode_eager_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_True_compile_mode_none_cpu_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_True_compile_mode_none_cpu_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_True_compile_mode_none_cuda_autograd_False, test/functorch/test_control_flow.py::TestControlFlow::test_scan_tuple_reverse_True_compile_mode_none_cuda_autograd_True, test/functorch/test_control_flow.py::TestControlFlow::test_scan_wrong_pytree, test/functorch/test_control_flow.py::TestControlFlow::test_scan_y_less_ndim_then_dim, test/functorch/test_control_flow.py::TestControlFlow::test_while_loop_gpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_eager_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_eager_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_eager_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_eager_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_eager_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_eager_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_eager_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_eager_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_none_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_none_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_none_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_none_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_none_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_none_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_none_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_binary_operator_compile_mode_none_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_combine_fn_wrong_meta_in_combine_fn, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_compile_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_compile_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_compile_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_compile_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_compile_dynamic_shape_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_compile_dynamic_shape_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_compile_dynamic_shape_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_compile_dynamic_shape_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_eager_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_eager_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_eager_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_eager_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_none_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_none_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_none_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_False_compile_mode_none_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_compile_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_compile_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_compile_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_compile_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_compile_dynamic_shape_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_compile_dynamic_shape_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_compile_dynamic_shape_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_compile_dynamic_shape_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_eager_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_eager_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_eager_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_eager_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_none_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_none_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_none_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_compile_reverse_True_compile_mode_none_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_eager_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_eager_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_eager_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_eager_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_eager_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_eager_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_eager_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_eager_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_none_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_none_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_none_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_none_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_none_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_none_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_none_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_complex_pytree_compile_mode_none_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_compile_dynamic_shape_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_compile_dynamic_shape_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_compile_dynamic_shape_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_compile_dynamic_shape_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_compile_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_compile_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_compile_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_compile_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_eager_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_eager_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_eager_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_eager_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_none_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_none_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_none_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_cond_in_combine_fn_compile_mode_none_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_compile_dynamic_shape_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_compile_dynamic_shape_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_compile_dynamic_shape_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_compile_dynamic_shape_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_compile_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_compile_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_compile_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_compile_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_eager_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_eager_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_eager_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_eager_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_none_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_none_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_none_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_compile_mode_none_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_different_input_size_wrong_dim, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_compile_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_compile_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_compile_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_compile_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_compile_dynamic_shape_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_compile_dynamic_shape_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_compile_dynamic_shape_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_compile_dynamic_shape_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_eager_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_eager_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_eager_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_eager_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_none_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_none_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_none_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_False_compile_mode_none_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_compile_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_compile_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_compile_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_compile_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_compile_dynamic_shape_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_compile_dynamic_shape_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_compile_dynamic_shape_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_compile_dynamic_shape_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_eager_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_eager_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_eager_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_eager_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_none_combine_mode_generic_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_none_combine_mode_generic_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_none_combine_mode_pointwise_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_reverse_True_compile_mode_none_combine_mode_pointwise_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_dim_shape_failure, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_compile_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_compile_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_compile_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_compile_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_eager_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_eager_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_eager_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_eager_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_none_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_none_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_none_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_generic_compile_mode_none_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_compile_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_compile_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_compile_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_compile_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_eager_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_eager_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_eager_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_eager_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_none_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_none_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_none_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_matmul_combine_mode_pointwise_compile_mode_none_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_compile_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_compile_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_compile_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_compile_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_eager_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_eager_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_eager_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_eager_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_none_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_none_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_none_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_generic_compile_mode_none_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_compile_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_compile_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_compile_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_compile_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_eager_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_eager_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_eager_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_eager_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_none_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_none_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_none_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_combine_mode_pointwise_compile_mode_none_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_first_False_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_first_False_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_first_False_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_first_False_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_first_True_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_first_True_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_first_True_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_dynamic_shape_reverse_first_True_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_reverse_first_False_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_reverse_first_False_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_reverse_first_False_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_reverse_first_False_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_reverse_first_True_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_reverse_first_True_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_reverse_first_True_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_compile_reverse_first_True_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_eager_reverse_first_False_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_eager_reverse_first_False_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_eager_reverse_first_False_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_eager_reverse_first_False_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_eager_reverse_first_True_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_eager_reverse_first_True_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_eager_reverse_first_True_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_eager_reverse_first_True_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_none_reverse_first_False_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_none_reverse_first_False_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_none_reverse_first_False_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_none_reverse_first_False_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_none_reverse_first_True_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_none_reverse_first_True_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_none_reverse_first_True_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_generic_compile_mode_none_reverse_first_True_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_first_False_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_first_False_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_first_False_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_first_False_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_first_True_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_first_True_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_first_True_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_dynamic_shape_reverse_first_True_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_reverse_first_False_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_reverse_first_False_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_reverse_first_False_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_reverse_first_False_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_reverse_first_True_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_reverse_first_True_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_reverse_first_True_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_compile_reverse_first_True_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_eager_reverse_first_False_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_eager_reverse_first_False_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_eager_reverse_first_False_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_eager_reverse_first_False_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_eager_reverse_first_True_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_eager_reverse_first_True_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_eager_reverse_first_True_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_eager_reverse_first_True_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_none_reverse_first_False_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_none_reverse_first_False_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_none_reverse_first_False_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_none_reverse_first_False_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_none_reverse_first_True_same_direction_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_none_reverse_first_True_same_direction_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_none_reverse_first_True_same_direction_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_downstream_scan_scan_different_dim_combine_mode_pointwise_compile_mode_none_reverse_first_True_same_direction_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_eager_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_eager_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_eager_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_eager_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_eager_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_eager_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_eager_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_eager_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_none_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_none_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_none_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_none_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_none_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_none_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_none_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_expand_in_combine_fn_compile_mode_none_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_eager_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_eager_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_eager_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_eager_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_eager_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_eager_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_eager_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_eager_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_none_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_none_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_none_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_none_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_none_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_none_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_none_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_compile_mode_none_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_compile_dynamic_shape_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_compile_dynamic_shape_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_compile_dynamic_shape_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_compile_dynamic_shape_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_compile_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_compile_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_compile_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_compile_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_eager_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_eager_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_eager_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_eager_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_none_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_none_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_none_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_fct_generic_compile_mode_none_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_eager_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_eager_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_eager_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_eager_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_eager_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_eager_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_eager_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_eager_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_none_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_none_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_none_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_none_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_none_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_none_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_none_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_nested_compile_mode_none_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_dynamic_shape_reverse_False_cpu_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_dynamic_shape_reverse_False_cpu_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_dynamic_shape_reverse_False_cuda_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_dynamic_shape_reverse_False_cuda_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_dynamic_shape_reverse_True_cpu_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_dynamic_shape_reverse_True_cpu_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_dynamic_shape_reverse_True_cuda_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_dynamic_shape_reverse_True_cuda_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_reverse_False_cpu_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_reverse_False_cpu_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_reverse_False_cuda_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_reverse_False_cuda_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_reverse_True_cpu_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_reverse_True_cpu_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_reverse_True_cuda_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_compile_reverse_True_cuda_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_eager_reverse_False_cpu_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_eager_reverse_False_cpu_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_eager_reverse_False_cuda_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_eager_reverse_False_cuda_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_eager_reverse_True_cpu_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_eager_reverse_True_cpu_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_eager_reverse_True_cuda_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_eager_reverse_True_cuda_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_none_reverse_False_cpu_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_none_reverse_False_cpu_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_none_reverse_False_cuda_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_none_reverse_False_cuda_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_none_reverse_True_cpu_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_none_reverse_True_cpu_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_none_reverse_True_cuda_combine_mode_generic, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_pytree_compile_mode_none_reverse_True_cuda_combine_mode_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_eager_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_eager_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_eager_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_eager_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_eager_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_eager_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_eager_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_eager_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_none_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_none_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_none_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_none_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_none_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_none_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_none_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_shape_check_compile_mode_none_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_eager_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_eager_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_eager_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_eager_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_eager_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_eager_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_eager_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_eager_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_none_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_none_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_none_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_none_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_none_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_none_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_none_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_freevars_simple_compile_mode_none_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_input_mutation, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_input_output_alias, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_compile_dynamic_shape_loop_type_for_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_compile_dynamic_shape_loop_type_for_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_compile_dynamic_shape_loop_type_for_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_compile_dynamic_shape_loop_type_for_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_compile_loop_type_for_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_compile_loop_type_for_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_compile_loop_type_for_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_compile_loop_type_for_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_eager_loop_type_for_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_eager_loop_type_for_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_eager_loop_type_for_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_eager_loop_type_for_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_none_loop_type_for_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_none_loop_type_for_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_none_loop_type_for_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_compile_mode_none_loop_type_for_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_loop_in_combine_fn_failure, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_map_in_combine_fn, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_nested, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_compile_dynamic_shape_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_compile_dynamic_shape_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_compile_dynamic_shape_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_compile_dynamic_shape_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_compile_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_compile_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_compile_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_compile_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_eager_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_eager_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_eager_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_eager_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_none_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_none_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_none_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_contiguous_tensor_compile_mode_none_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_False_compile_mode_compile_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_False_compile_mode_compile_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_False_compile_mode_compile_dynamic_shape_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_False_compile_mode_compile_dynamic_shape_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_False_compile_mode_eager_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_False_compile_mode_eager_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_False_compile_mode_none_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_False_compile_mode_none_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_True_compile_mode_compile_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_True_compile_mode_compile_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_True_compile_mode_compile_dynamic_shape_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_True_compile_mode_compile_dynamic_shape_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_True_compile_mode_eager_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_True_compile_mode_eager_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_True_compile_mode_none_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_non_pointwise_generic_reverse_True_compile_mode_none_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_output_output_alias, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_pytree_output, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_sparse_tensor, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_dynamic_shape_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_compile_dynamic_shape_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_eager_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_eager_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_eager_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_eager_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_eager_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_eager_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_eager_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_eager_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_none_combine_mode_generic_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_none_combine_mode_generic_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_none_combine_mode_generic_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_none_combine_mode_generic_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_none_combine_mode_pointwise_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_none_combine_mode_pointwise_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_none_combine_mode_pointwise_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_tuple_compile_mode_none_combine_mode_pointwise_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_compile_dynamic_shape_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_compile_dynamic_shape_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_compile_dynamic_shape_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_compile_dynamic_shape_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_compile_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_compile_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_compile_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_compile_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_eager_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_eager_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_eager_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_eager_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_none_reverse_False_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_none_reverse_False_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_none_reverse_True_cpu, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_vmap_in_combine_fn_compile_mode_none_reverse_True_cuda, test/functorch/test_control_flow.py::AssociativeScanTests::test_associative_scan_wrong_pytree, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_accepts_torch_function_as_inputs, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_autograd_backward, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_eager_run_with_item, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_aot_func_check_functional, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_data_dependent_pred, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_input_aliasing_with_aot_func, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_input_mutation_on_false_branch, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_input_mutation_on_true_branch, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_nested, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_nested_input_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_nested_input_mutation_with_aot_func, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_functionalized_output_alias_input, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_make_fx_preserve_stack_trace_for_nodes_in_subgraph, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_merge_graph_preserves_ph_meta, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_mismatched_branch_output_dynamic_False_backend_aot_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_mismatched_branch_output_dynamic_False_backend_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_mismatched_branch_output_dynamic_True_backend_aot_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_mismatched_branch_output_dynamic_True_backend_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_mismatched_branch_strided_output_dynamic_False_backend_aot_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_mismatched_branch_strided_output_dynamic_False_backend_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_mismatched_branch_strided_output_dynamic_True_backend_aot_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_mismatched_branch_strided_output_dynamic_True_backend_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_fake_tensor, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_multi, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_multi_fake_tensor, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_other_inputs, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_traced_other_inputs_fake_tensor, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_with_closure, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_nested_with_closure_graph_module, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_no_dynamo_cache_limit, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_retrace_functionalized, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_simple_with_linear_compile_check_graph, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_subgraph_same_shape_env_as_parent, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_symint_closure, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_symint_operands_requires_grad_False, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_symint_operands_requires_grad_True, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_trace_set__and_mutate_input, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_trace_set__and_mutate_intermediate, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_traced_not_nested, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_traced_not_nested_fake_tensor, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_function_nOperands_0_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_function_nOperands_0_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_function_nOperands_0_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_function_nOperands_0_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_function_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_function_nOperands_1_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_function_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_function_nOperands_1_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_module_nOperands_0_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_module_nOperands_0_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_module_nOperands_0_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_module_nOperands_0_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_module_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_module_nOperands_1_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_module_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_module_nOperands_1_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_object_nOperands_0_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_object_nOperands_0_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_object_nOperands_0_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_object_nOperands_0_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_object_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_object_nOperands_1_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_object_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_boolTensor_innerFnType_object_nOperands_1_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_function_nOperands_0_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_function_nOperands_0_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_function_nOperands_0_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_function_nOperands_0_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_function_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_function_nOperands_1_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_function_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_function_nOperands_1_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_module_nOperands_0_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_module_nOperands_0_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_module_nOperands_0_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_module_nOperands_0_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_module_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_module_nOperands_1_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_module_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_module_nOperands_1_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_object_nOperands_0_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_object_nOperands_0_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_object_nOperands_0_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_object_nOperands_0_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_object_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_object_nOperands_1_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_object_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_bool_innerFnType_object_nOperands_1_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_function_nOperands_0_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_function_nOperands_0_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_function_nOperands_0_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_function_nOperands_0_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_function_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_function_nOperands_1_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_function_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_function_nOperands_1_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_module_nOperands_0_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_module_nOperands_0_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_module_nOperands_0_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_module_nOperands_0_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_module_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_module_nOperands_1_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_module_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_module_nOperands_1_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_object_nOperands_0_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_object_nOperands_0_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_object_nOperands_0_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_object_nOperands_0_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_object_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_object_nOperands_1_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_object_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_floatTensor_innerFnType_object_nOperands_1_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_function_nOperands_0_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_function_nOperands_0_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_function_nOperands_0_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_function_nOperands_0_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_function_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_function_nOperands_1_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_function_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_function_nOperands_1_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_module_nOperands_0_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_module_nOperands_0_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_module_nOperands_0_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_module_nOperands_0_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_module_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_module_nOperands_1_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_module_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_module_nOperands_1_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_object_nOperands_0_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_object_nOperands_0_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_object_nOperands_0_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_object_nOperands_0_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_object_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_object_nOperands_1_nClosure_0_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_object_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_tracing_with_valid_inputs_predType_intTensor_innerFnType_object_nOperands_1_nClosure_1_nesting_2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_unbacked_symint_closure, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_multiple_args_with_closure, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_multiple_inputs, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_multiple_outputs_nClosure_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_multiple_outputs_nClosure_1, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_predType_boolTensor_innerFnType_function_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_predType_boolTensor_innerFnType_function_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_predType_boolTensor_innerFnType_function_nOperands_2_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_predType_boolTensor_innerFnType_function_nOperands_2_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_predType_boolTensor_innerFnType_module_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_predType_boolTensor_innerFnType_module_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_predType_boolTensor_innerFnType_module_nOperands_2_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_predType_boolTensor_innerFnType_module_nOperands_2_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_predType_boolTensor_innerFnType_object_nOperands_1_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_predType_boolTensor_innerFnType_object_nOperands_1_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_predType_boolTensor_innerFnType_object_nOperands_2_nClosure_0_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_predType_boolTensor_innerFnType_object_nOperands_2_nClosure_1_nesting_0, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_simple, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_vmap_single_input_with_closure, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_with_consecutive_make_fx_symbolic, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_with_module_param_closure, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_with_module_python_scalar_closure, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_with_sym_pred, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_with_tensor_closure, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_with_tensor_closure_graph_module, test/functorch/test_control_flow.py::TestControlFlowTraced::test_cond_with_unbacked_sym_pred, test/functorch/test_control_flow.py::TestControlFlowTraced::test_hop_raises_if_not_overriding_call, test/functorch/test_control_flow.py::TestControlFlowTraced::test_input_input_alias, test/functorch/test_control_flow.py::TestControlFlowTraced::test_input_mutation_inference_mode_False, test/functorch/test_control_flow.py::TestControlFlowTraced::test_input_mutation_inference_mode_True, test/functorch/test_control_flow.py::TestControlFlowTraced::test_input_output_alias, test/functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized, test/functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized_aot_func, test/functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized_arg_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized_elem_alias, test/functorch/test_control_flow.py::TestControlFlowTraced::test_map_functionalized_elem_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_map_unfunc_boolean_tensor_for_nested_map_cond, test/functorch/test_control_flow.py::TestControlFlowTraced::test_merge_output, test/functorch/test_control_flow.py::TestControlFlowTraced::test_nested_cond_map_cond_symbolic, test/functorch/test_control_flow.py::TestControlFlowTraced::test_nested_map_cond_real, test/functorch/test_control_flow.py::TestControlFlowTraced::test_nested_map_cond_symbolic, test/functorch/test_control_flow.py::TestControlFlowTraced::test_raise_error_on_mismatch_tensor_size, test/functorch/test_control_flow.py::TestControlFlowTraced::test_raise_error_on_mismatch_tensor_size_fake_tensor, test/functorch/test_control_flow.py::TestControlFlowTraced::test_raise_error_on_mismatch_type_size, test/functorch/test_control_flow.py::TestControlFlowTraced::test_raise_error_on_mismatch_type_size_fake_tensor, test/functorch/test_control_flow.py::TestControlFlowTraced::test_scan_functionalized, test/functorch/test_control_flow.py::TestControlFlowTraced::test_scan_functionalized_elem_alias, test/functorch/test_control_flow.py::TestControlFlowTraced::test_scan_functionalized_elem_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_scan_pytree_closure, test/functorch/test_control_flow.py::TestControlFlowTraced::test_tracing_map_autograd_aot_functionalized, test/functorch/test_control_flow.py::TestControlFlowTraced::test_tracing_map_autograd_symbolic_dict, test/functorch/test_control_flow.py::TestControlFlowTraced::test_tracing_map_autograd_symbolic_list, test/functorch/test_control_flow.py::TestControlFlowTraced::test_tracing_map_autograd_symbolic_simple, test/functorch/test_control_flow.py::TestControlFlowTraced::test_tracing_map_real, test/functorch/test_control_flow.py::TestControlFlowTraced::test_tracing_map_symbolic_dict, test/functorch/test_control_flow.py::TestControlFlowTraced::test_tracing_map_symbolic_list, test/functorch/test_control_flow.py::TestControlFlowTraced::test_tracing_map_symbolic_simple, test/functorch/test_control_flow.py::TestControlFlowTraced::test_two_hops_not_sharing_code_obj, test/functorch/test_control_flow.py::TestControlFlowTraced::test_vmap_vmap_boolcond_False, test/functorch/test_control_flow.py::TestControlFlowTraced::test_vmap_vmap_boolcond_True, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_aot_eager_while_loop_test_const_and_symint_output, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_aot_eager_while_loop_test_int_carry, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_aot_eager_while_loop_test_nested, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_aot_eager_while_loop_test_nested2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_aot_eager_while_loop_test_nested_with_linear, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_aot_eager_while_loop_test_pytree_int_carry, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_aot_eager_while_loop_test_simple, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_aot_eager_while_loop_test_simple_with_linear, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_aot_eager_while_loop_test_simple_with_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_aot_eager_while_loop_test_simple_with_pytree_carry, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_eager_while_loop_test_const_and_symint_output, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_eager_while_loop_test_int_carry, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_eager_while_loop_test_nested, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_eager_while_loop_test_nested2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_eager_while_loop_test_nested_with_linear, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_eager_while_loop_test_pytree_int_carry, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_eager_while_loop_test_simple, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_eager_while_loop_test_simple_with_linear, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_eager_while_loop_test_simple_with_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_compile_backend_eager_while_loop_test_simple_with_pytree_carry, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_cpp_while_loop_test_nested, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_cpp_while_loop_test_nested2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_cpp_while_loop_test_simple, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_cpp_while_loop_test_simple_with_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_cpp_while_loop_test_simple_with_pytree_carry, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_functorch_while_loop_test_nested, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_functorch_while_loop_test_nested2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_functorch_while_loop_test_simple, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_functorch_while_loop_test_simple_with_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_functorch_while_loop_test_simple_with_pytree_carry, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_no_while_loop_test_nested, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_no_while_loop_test_nested2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_no_while_loop_test_simple, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_no_while_loop_test_simple_with_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_no_while_loop_test_simple_with_pytree_carry, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_python_while_loop_test_nested, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_python_while_loop_test_nested2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_python_while_loop_test_simple, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_python_while_loop_test_simple_with_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_functionalize_func_type_python_while_loop_test_simple_with_pytree_carry, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_nested2_traced, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_nested_traced, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_constant_and_symint_output_compile_dynamic_False_backend_aot_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_constant_and_symint_output_compile_dynamic_False_backend_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_constant_and_symint_output_compile_dynamic_True_backend_aot_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_constant_and_symint_output_compile_dynamic_True_backend_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_constant_and_symint_output_export_strict_False_dynamic_False, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_constant_and_symint_output_export_strict_False_dynamic_True, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_constant_and_symint_output_export_strict_True_dynamic_False, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_constant_and_symint_output_export_strict_True_dynamic_True, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_int_carry_compile_dynamic_False_backend_aot_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_int_carry_compile_dynamic_False_backend_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_int_carry_compile_dynamic_True_backend_aot_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_int_carry_compile_dynamic_True_backend_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_int_carry_export_strict_False_dynamic_False, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_int_carry_export_strict_False_dynamic_True, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_mismatch_in_meta, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_pytree_int_carry_compile_dynamic_False_backend_aot_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_pytree_int_carry_compile_dynamic_False_backend_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_pytree_int_carry_compile_dynamic_True_backend_aot_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_pytree_int_carry_compile_dynamic_True_backend_eager, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_pytree_int_carry_export_strict_False_dynamic_False, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_pytree_int_carry_export_strict_False_dynamic_True, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_pytree_int_carry_export_strict_True_dynamic_False, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_op_pytree_int_carry_export_strict_True_dynamic_True, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_pytree_carry, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_simple_functionalize_check_graph_func_type_cpp, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_simple_functionalize_check_graph_func_type_functorch, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_simple_functionalize_check_graph_func_type_no, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_simple_functionalize_check_graph_func_type_python, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_simple_with_linear_compile_check_graph, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_tracing_while_loop_test_nested, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_tracing_while_loop_test_nested2, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_tracing_while_loop_test_nested_with_linear, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_tracing_while_loop_test_simple, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_tracing_while_loop_test_simple_with_linear, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_tracing_while_loop_test_simple_with_mutation, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_tracing_while_loop_test_simple_with_pytree_carry, test/functorch/test_control_flow.py::TestControlFlowTraced::test_while_loop_unbacked_bindings, test/functorch/test_control_flow.py::TestHopSchema::test_function_schema_gen, test/functorch/test_control_flow.py::TestHopSchema::test_list_gen_schema_type_GraphModule, test/functorch/test_control_flow.py::TestHopSchema::test_list_gen_schema_type_ScriptObj, test/functorch/test_control_flow.py::TestHopSchema::test_list_gen_schema_type_SymBool, test/functorch/test_control_flow.py::TestHopSchema::test_list_gen_schema_type_SymInt, test/functorch/test_control_flow.py::TestHopSchema::test_list_gen_schema_type_Tensor, test/functorch/test_control_flow.py::TestHopSchema::test_list_gen_schema_type_bool, test/functorch/test_control_flow.py::TestHopSchema::test_list_gen_schema_type_float, test/functorch/test_control_flow.py::TestHopSchema::test_list_gen_schema_type_int, test/functorch/test_control_flow.py::TestHopSchema::test_list_gen_schema_type_str, test/functorch/test_control_flow.py::TestHopSchema::test_schema_tree_spec, test/functorch/test_control_flow.py::TestHopSchema::test_type_gen_schema_type_GraphModule, test/functorch/test_control_flow.py::TestHopSchema::test_type_gen_schema_type_ScriptObj, test/functorch/test_control_flow.py::TestHopSchema::test_type_gen_schema_type_SymBool, test/functorch/test_control_flow.py::TestHopSchema::test_type_gen_schema_type_SymInt, test/functorch/test_control_flow.py::TestHopSchema::test_type_gen_schema_type_Tensor, test/functorch/test_control_flow.py::TestHopSchema::test_type_gen_schema_type_bool, test/functorch/test_control_flow.py::TestHopSchema::test_type_gen_schema_type_float, test/functorch/test_control_flow.py::TestHopSchema::test_type_gen_schema_type_int, test/functorch/test_control_flow.py::TestHopSchema::test_type_gen_schema_type_str, test/functorch/test_control_flow.py::TestHopSchema::test_while_loop_schema_gen 2025-07-17T09:54:17.2871798Z 2025-07-17T09:54:17.2871996Z Running inductor/test_b2b_gemm 1/1 ... [2025-07-17 09:54:17.199835] 2025-07-17T09:54:17.2872299Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:54:17.2873368Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_b2b_gemm.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:54:17.200178] 2025-07-17T09:54:24.5279038Z 2025-07-17T09:54:24.5279978Z inductor/test_b2b_gemm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_b2b_gemm_1.1_c372f4d3d2221228_.log 2025-07-17T09:54:24.5280516Z Running 0 items in this shard: 2025-07-17T09:54:24.5280652Z 2025-07-17T09:54:24.5280866Z Running inductor/test_compile_subprocess 2/2 ... [2025-07-17 09:54:24.527803] 2025-07-17T09:54:24.5281190Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:54:24.5282972Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_compile_subprocess.py', '-m', 'not serial', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:54:24.528126] 2025-07-17T09:58:22.8942164Z 2025-07-17T09:58:22.8943338Z inductor/test_aot_inductor_package 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_package_1.1_9c7295f0efb748f3_.log 2025-07-17T09:58:22.8977783Z Running 80 items in this shard: test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_add, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_bool_input, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_compile_after_package, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_compile_after_package_multi_arch, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_compile_after_package_static, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_compile_with_exporter, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_deepcopy_compiled_model, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_duplicate_calls, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_linear, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_loading_wrong_model, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_metadata, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_multiple_methods, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_package_shared_weights, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_package_user_managed_weight, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_package_weights_on_disk_nested_module, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_package_without_weight, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_remove_intermediate_files, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_save_buffer, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_specified_output_dir, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cpu::test_update_weights, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_add, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_bool_input, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_compile_after_package, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_compile_after_package_multi_arch, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_compile_after_package_static, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_compile_with_exporter, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_deepcopy_compiled_model, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_duplicate_calls, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_linear, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_loading_wrong_model, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_metadata, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_multiple_methods, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_package_shared_weights, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_package_user_managed_weight, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_package_weights_on_disk_nested_module, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_package_without_weight, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_remove_intermediate_files, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_save_buffer, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_specified_output_dir, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cpu::test_update_weights, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_add, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_bool_input, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_compile_after_package, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_compile_after_package_multi_arch, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_compile_after_package_static, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_compile_with_exporter, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_deepcopy_compiled_model, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_duplicate_calls, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_linear, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_loading_wrong_model, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_metadata, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_multiple_methods, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_package_shared_weights, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_package_user_managed_weight, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_package_weights_on_disk_nested_module, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_package_without_weight, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_remove_intermediate_files, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_save_buffer, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_specified_output_dir, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackage_cuda::test_update_weights, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_add, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_bool_input, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_compile_after_package, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_compile_after_package_multi_arch, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_compile_after_package_static, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_compile_with_exporter, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_deepcopy_compiled_model, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_duplicate_calls, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_linear, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_loading_wrong_model, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_metadata, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_multiple_methods, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_package_shared_weights, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_package_user_managed_weight, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_package_weights_on_disk_nested_module, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_package_without_weight, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_remove_intermediate_files, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_save_buffer, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_specified_output_dir, test/inductor/test_aot_inductor_package.py::TestAOTInductorPackageCpp_cuda::test_update_weights 2025-07-17T09:58:22.8997206Z 2025-07-17T09:58:22.8997379Z Running inductor/test_compiled_optimizers 2/3 ... [2025-07-17 09:58:22.898937] 2025-07-17T09:58:22.8997698Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T09:58:22.8998476Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_compiled_optimizers.py', '-m', 'not serial', '--shard-id=2', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 09:58:22.899239] 2025-07-17T10:01:46.5913477Z 2025-07-17T10:01:46.5914599Z inductor/test_compile_subprocess 2/2 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_compile_subprocess_2.2_e4de1fdd3a3f5a49_.log 2025-07-17T10:01:46.6001287Z Running 427 items in this shard: test/inductor/test_compile_subprocess.py::TestSubprocess::test_async, test/inductor/test_compile_subprocess.py::GPUTests::test__dyn_quant_matmul_4bit_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test__unsafe_masked_index_put_accumulate_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_adaptive_avg_pool2d1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_adaptive_avg_pool2d_low_prec_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_adaptive_max_pool2d2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_add_complex4_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_add_complex5_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_add_complex6_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_add_complex_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_add_const_float_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_add_const_int_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_aliased_buffer_reuse_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_any_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_aoti_eager_cache_hit_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_aoti_eager_dtype_device_layout_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_aoti_eager_support_str_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_aoti_eager_with_persistent_cache_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_arange1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_arange2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_arange3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_arange5_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_arange6_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_argmax_argmin_with_duplicates_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_argmax_argmin_with_nan_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_argmax_min_int32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_as_strided_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_assert_alignment_op_name_pass_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_assert_size_stride_op_name_fail_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_async, test/inductor/test_compile_subprocess.py::GPUTests::test_avg_pool2d2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_avg_pool2d3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_avg_pool2d7_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_avg_pool2d_backward2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_avg_pool2d_backward_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_avg_pool3d_backward_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_avg_pool_errors_with_uint_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_batch_norm_2d_2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_batch_norm_2d_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bfloat16_to_int16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bitwise3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bmm2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_add_autotune_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_default_kwargs_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int16_int16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int16_int32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int16_int64_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int16_int8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int32_int64_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int32_uint8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int64_int16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int64_int32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int64_uint8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int8_int16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int8_int32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int8_int64_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int8_int8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_int8_uint8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_uint8_int32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_uint8_int64_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_int_uint8_uint8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_bucketize_nd_tiling_False_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_buffer_batch_norm_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_buffer_copied_in_graph_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_builtins_round_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_builtins_round_float_ndigits_pos_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cat_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cat_empty_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cat_empty_index_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cat_extern_kernel_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cat_of_loops_and_extern_kernel_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cat_uint8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cat_unbacked_empty_1d_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cat_upcasting_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cauchy_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_check_stack_no_cycles_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_clamp_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_clamp_type_promotion_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_clone_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_computed_buffer_inlining_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_concat_add_inplace_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_config_option_dont_assume_alignment_cudagraphs_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_config_option_dont_assume_alignment_recompiles_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_consecutive_split_cumprod_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_consecutive_split_cumsum_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_constant_pad_2d_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_constant_pad_3d_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_constant_pad_nd_inplace_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_conv2d_backward_channels_last_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_conv2d_channels_last_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_conv3d_channels_last_use_block_ptr_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_conv_functional_bn_fuse_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_convolution2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_convolution5_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cos_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cudnn_rnn_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cummin_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cumprod_zero_dim_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cumsum_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cumsum_no_mask_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_cumsum_zero_dim_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_custom_op_3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_custom_scan_op_compiled_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_custom_scan_would_split_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_deterministic_codegen_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_deterministic_codegen_on_graph_break_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_device_assert_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dist_bf16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dist_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_div1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_div5_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_div6_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_div7_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_div_zero_dim_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dropout2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dropout3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dropout_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dropout_deterministic_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dropout_trivial_0_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_bfloat16_float16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_bfloat16_float32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_bfloat16_float64_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_bfloat16_int16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_bfloat16_int8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_bfloat16_uint8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float16_float16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float16_float32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float16_float64_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float16_int16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float16_int32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float16_int64_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float16_int8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float16_uint8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float32_bfloat16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float32_float32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float32_int16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float32_int32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float32_uint8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float64_float16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float64_int32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float64_int64_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_float64_int8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int16_float32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int16_int32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int16_int8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int16_uint8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int32_bfloat16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int32_float32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int32_int16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int32_int32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int32_int8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int32_uint8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int64_bfloat16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int64_float16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int64_int16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int64_int32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int8_float32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int8_int16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_int8_int8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_uint8_float16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_uint8_float32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_uint8_float64_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_dtypeview_uint8_int64_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_embedding_bag_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_embedding_sparse_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_empty1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_empty2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_erfc_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_erfinv_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_expanded_reduction_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_expm1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fallback_mutable_op_basic_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fallback_mutable_op_list_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fallback_mutable_op_list_tensor_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fallback_mutable_op_no_mutated_tensors_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fallback_mutable_op_with_return_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fft_real_input_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fill1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_float16_to_int16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_float32_to_int32_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_float_index_expression_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_float_index_expression_type_promotion_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fmin_fmax_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fmod_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fmod_zero_dim_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fractional_max_pool2d1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fractional_max_pool2d2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fractional_max_pool2d4_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_full_truncation_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fuse_large_params_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fuse_tiled_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_fusing_write_into_disjoint_read_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_gather2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_gather3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_generate_rand_fp8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_generated_code_has_alignment_assert_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_generated_code_has_size_stride_assert_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_glu_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_graph_partition_arange1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_graph_partition_argmax_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_graph_partition_both_scalars_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_graph_partition_constant_tensor1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_graph_partition_constant_tensor2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_graph_partition_no_inputs_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_graph_partition_pad_dynamic_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_graph_partition_refcount_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_graph_partition_unbacked_symint_as_output_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_grid_sampler_2d_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_hardsigmoid_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_hardswish_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_horizonal_fusion2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_dynamic_shapes_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_propagation_abs_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_propagation_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_propagation_device_assert_masked_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_propagation_flip_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_propagation_floordiv_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_propagation_nested_indirect_indexing_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_propagation_remainder_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_put2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_put_fallback1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_put_fallback2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_select_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_index_tensor_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_inductor_assert_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_inductor_layout_optimization_input_mutations_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_inductor_multiple_specializations_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_inner_fn_str_and_stride_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_inplace_activations_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_inplace_mixed_dtype_ops_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_inplace_where_pointwise_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_input_mutation1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_input_mutation3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_input_mutation4_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_int_input_dynamic_shapes_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_invalid_operand_issue1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_isinf2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_isinf_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_large_grid_use_block_ptr_False_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_large_grid_use_block_ptr_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_large_pointwise_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_large_strided_reduction_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_layer_norm_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_leaky_relu_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_lerp_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_lgamma_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_like_channels_last_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_like_rands2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_like_rands3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_like_rands_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_linear1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_linear2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_linear_dynamic_maxautotune_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_linear_float64_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_linear_mixed_dtype_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_linspace1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_linspace3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_log1p_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_log_softmax_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_long_tensor_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_low_memory_max_pool_dilation_1_dim_2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_masked_fill_promotion_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_matmul_layer_norm_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_max_min_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_max_pool2d2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_max_pool2d6_dilation_2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_max_pool2d7_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_max_pool2d8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_max_pool2d_with_indices_backward5_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_min_max_reduction_nan_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_mixed_mm2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_mm_mixed_dtype_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_mm_views_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_multi_device_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_multi_gpu_device_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_multilayer_any_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_multilayer_sum_low_prec_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_multilayer_var_lowp_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_mutable_custom_op_fixed_layout2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_mutable_custom_op_fixed_layout_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_mutations_loop_fusion_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_nan_to_num_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_narrow_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_neg_max_uint8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_nll_loss_backward_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_nll_loss_forward_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_no_specization_over_symbolic_value_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_nonzero_unbacked_refinement_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_output_strides_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pad_cast_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pad_view_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pattern_matcher_multi_user_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_permute2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pixel_shuffle_channels_last_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_airy_ai_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_chebyshev_polynomial_v_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_chebyshev_polynomial_w_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_erf_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_erfcx_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_expit_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_gammaincc_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_hermite_polynomial_he_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_i0_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_i0e_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_i1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_i1e_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_log_ndtr_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_logit_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_multigammaln_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_ndtr_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_psi_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_shifted_chebyshev_polynomial_t_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_shifted_chebyshev_polynomial_u_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_shifted_chebyshev_polynomial_v_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pointwise_shifted_chebyshev_polynomial_w_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_polar_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pow1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pow2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_pow_by_natural_log2_dynamic_shapes_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_prepare_softmax_with_fast_math_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_randint_distribution_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_randint_int64_mod_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_randint_kernel_count_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_randn_like_empty_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_randn_with_dtype_and_device_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_reduction2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_reduction4_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_reduction5_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remainder_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_clone_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_copy_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_repeat_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_replication_pad_errors_with_bool_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_require_stride_expanded_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_resize_as_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_resize_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_roi_align_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_roll_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_round_correctness_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_round_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_scalar_cpu_tensor_arg_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_scalar_input_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_scalar_output_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_scaled_dot_product_attention_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_scaled_dot_product_efficient_attention_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_scatter1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_scatter4_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_scatter5_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_scatter_add2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_scatter_add3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_scatter_reduce2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_scheduler_vertical_fusion1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sdpa_prefer_nd_tiling_False_use_block_ptr_False_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sdpa_prefer_nd_tiling_False_use_block_ptr_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sdpa_prefer_nd_tiling_True_use_block_ptr_False_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sdpa_unaligned_mask_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sdpa_unaligned_mask_freezing_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_searchsorted_broadcast_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_shape_prop_torch_ones_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_should_pad_bench_for_bmm_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sigmoid_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sign_dtype_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_signbit_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_simplify_loops_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sin_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_single_elem_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_single_elem_indirect_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sizehint_issue1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_slice1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_slice3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_slice_mutation1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_slice_mutation2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_slice_scatter2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_slice_scatter5_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_slice_scatter_reinplace_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_softmax_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_softmax_one_kernel_persist_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_split_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_split_cumprod_low_prec_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_split_cumsum_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_split_cumsum_index_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_split_reduction_dynamic_shape_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_split_reduction_with_int64_size_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_split_with_integer_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_split_with_unbacked_symints_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sqrt_dynamic_shapes_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_squeeze1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_squeeze2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_squeeze_varargs_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_stack_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sum2_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sum4_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_sum5_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_tan_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_tensor_index_slice_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_to_device_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_to_dtype_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_transpose_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_uint_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_unbind_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_unfold_zero_dimension_tensor_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_unspec_inputs_bfloat16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_unspec_inputs_float16_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_unspec_inputs_uint8_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_unsqueeze_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_upsample_bicubic2d_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_upsample_bilinear2d_b_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_upsample_nearest2d_backward_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_upsample_nearest2d_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_upsample_nearest3d_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_div_by_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_vectorized_ops_masked_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_vectorized_ops_masked_var_novec_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_vertical_fusion1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_view_on_aliased_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_view_uint8_through_differing_bitwidths_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_views1_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_views3_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_views7_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_xblock_divides_xnumel_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_zeros_cuda 2025-07-17T10:01:46.6082104Z 2025-07-17T10:01:46.6082341Z Running inductor/test_cuda_repro 1/1 ... [2025-07-17 10:01:46.593581] 2025-07-17T10:01:46.6082765Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:01:46.6083564Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_cuda_repro.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:01:46.593931] 2025-07-17T10:03:34.5880185Z 2025-07-17T10:03:34.5881169Z inductor/test_cuda_repro 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cuda_repro_1.1_c7b505d7605eb5e5_.log 2025-07-17T10:03:34.5896523Z Running 78 items in this shard: test/inductor/test_cuda_repro.py::CudaReproTests::test_3d_tiling, test/inductor/test_cuda_repro.py::CudaReproTests::test_accuracy_issue1, test/inductor/test_cuda_repro.py::CudaReproTests::test_adaptive_avg_pool3d_issue_157248, test/inductor/test_cuda_repro.py::CudaReproTests::test_atomic_add_bfloat16, test/inductor/test_cuda_repro.py::CudaReproTests::test_atomic_add_bfloat16_config, test/inductor/test_cuda_repro.py::CudaReproTests::test_autotune_inplace_kernel, test/inductor/test_cuda_repro.py::CudaReproTests::test_backward_context, test/inductor/test_cuda_repro.py::CudaReproTests::test_bool_emulate_low_precision, test/inductor/test_cuda_repro.py::CudaReproTests::test_bucketize_dynamic_dense, test/inductor/test_cuda_repro.py::CudaReproTests::test_bucketize_epilogue, test/inductor/test_cuda_repro.py::CudaReproTests::test_cat_int8_one_kernel, test/inductor/test_cuda_repro.py::CudaReproTests::test_cpu_index, test/inductor/test_cuda_repro.py::CudaReproTests::test_deterministic_algorithms, test/inductor/test_cuda_repro.py::CudaReproTests::test_dont_inplace_disjoint_accesses, test/inductor/test_cuda_repro.py::CudaReproTests::test_dtype_factory_issue, test/inductor/test_cuda_repro.py::CudaReproTests::test_dynamic_persistent_reductions, test/inductor/test_cuda_repro.py::CudaReproTests::test_dynamic_shapes, test/inductor/test_cuda_repro.py::CudaReproTests::test_dynamic_to_static_cudagraphs, test/inductor/test_cuda_repro.py::CudaReproTests::test_effn_attn_bias_padding, test/inductor/test_cuda_repro.py::CudaReproTests::test_effn_attn_bias_padding_misaligned, test/inductor/test_cuda_repro.py::CudaReproTests::test_embedding_var_mean, test/inductor/test_cuda_repro.py::CudaReproTests::test_emulate_low_precision, test/inductor/test_cuda_repro.py::CudaReproTests::test_epilogue_fusion_with_view, test/inductor/test_cuda_repro.py::CudaReproTests::test_expanded_inputs_cudagraphs, test/inductor/test_cuda_repro.py::CudaReproTests::test_expanded_inputs_cudagraphs_no_size_asserts, test/inductor/test_cuda_repro.py::CudaReproTests::test_flash_attention_dynamic, test/inductor/test_cuda_repro.py::CudaReproTests::test_float64_constants, test/inductor/test_cuda_repro.py::CudaReproTests::test_float8_e8m0fnu, test/inductor/test_cuda_repro.py::CudaReproTests::test_full_copy, test/inductor/test_cuda_repro.py::CudaReproTests::test_index_add_fallback, test/inductor/test_cuda_repro.py::CudaReproTests::test_index_put_cudagraph, test/inductor/test_cuda_repro.py::CudaReproTests::test_index_put_inplace_cudagraph, test/inductor/test_cuda_repro.py::CudaReproTests::test_index_put_issue, test/inductor/test_cuda_repro.py::CudaReproTests::test_index_put_no_fallback_cudagraph, test/inductor/test_cuda_repro.py::CudaReproTests::test_indirect_indexing_dense_mask, test/inductor/test_cuda_repro.py::CudaReproTests::test_inductor_output_aliases_intermediate, test/inductor/test_cuda_repro.py::CudaReproTests::test_inplace_add_alpha_autotune, test/inductor/test_cuda_repro.py::CudaReproTests::test_inplace_buffer_autotune, test/inductor/test_cuda_repro.py::CudaReproTests::test_inplace_updates_cudagraphs, test/inductor/test_cuda_repro.py::CudaReproTests::test_input_channels_last, test/inductor/test_cuda_repro.py::CudaReproTests::test_int64_index_intermediate, test/inductor/test_cuda_repro.py::CudaReproTests::test_issue100806, test/inductor/test_cuda_repro.py::CudaReproTests::test_issue103461, test/inductor/test_cuda_repro.py::CudaReproTests::test_issue103481, test/inductor/test_cuda_repro.py::CudaReproTests::test_issue104759, test/inductor/test_cuda_repro.py::CudaReproTests::test_issue97695_1input, test/inductor/test_cuda_repro.py::CudaReproTests::test_issue97695_2input, test/inductor/test_cuda_repro.py::CudaReproTests::test_issue_103924, test/inductor/test_cuda_repro.py::CudaReproTests::test_libdevice_routing, test/inductor/test_cuda_repro.py::CudaReproTests::test_linear_cpu_input, test/inductor/test_cuda_repro.py::CudaReproTests::test_linear_with_zero_infeature_size, test/inductor/test_cuda_repro.py::CudaReproTests::test_lookup_seed_backward, test/inductor/test_cuda_repro.py::CudaReproTests::test_max_autotune_nograd, test/inductor/test_cuda_repro.py::CudaReproTests::test_memory_history_inductor, test/inductor/test_cuda_repro.py::CudaReproTests::test_multi_output_layout_fallback, test/inductor/test_cuda_repro.py::CudaReproTests::test_mutated_aligned_tensor, test/inductor/test_cuda_repro.py::CudaReproTests::test_negative_arange_dynamic_shapes, test/inductor/test_cuda_repro.py::CudaReproTests::test_no_device_idx_repro_cudagraphs, test/inductor/test_cuda_repro.py::CudaReproTests::test_non_commutative_scan_op, test/inductor/test_cuda_repro.py::CudaReproTests::test_non_contiguous_unaligned_input_indices, test/inductor/test_cuda_repro.py::CudaReproTests::test_not_initializing_wrong_device, test/inductor/test_cuda_repro.py::CudaReproTests::test_permute_fusion, test/inductor/test_cuda_repro.py::CudaReproTests::test_reflection_pad_loop_order, test/inductor/test_cuda_repro.py::CudaReproTests::test_repeated_masked_load, test/inductor/test_cuda_repro.py::CudaReproTests::test_scalar_triton_index, test/inductor/test_cuda_repro.py::CudaReproTests::test_scaled_dot_product_efficient_attention_backward, test/inductor/test_cuda_repro.py::CudaReproTests::test_scatter_index_not_wrapped, test/inductor/test_cuda_repro.py::CudaReproTests::test_selecsls42b_misaligned_address, test/inductor/test_cuda_repro.py::CudaReproTests::test_simplify_dims, test/inductor/test_cuda_repro.py::CudaReproTests::test_sort_stride_issue, test/inductor/test_cuda_repro.py::CudaReproTests::test_sorted_masks, test/inductor/test_cuda_repro.py::CudaReproTests::test_split_reduction_channels_last, test/inductor/test_cuda_repro.py::CudaReproTests::test_split_reduction_transposed, test/inductor/test_cuda_repro.py::CudaReproTests::test_triton_interpret, test/inductor/test_cuda_repro.py::CudaReproTests::test_uint_view_copy, test/inductor/test_cuda_repro.py::CudaReproTests::test_unspec_inputs_interop, test/inductor/test_cuda_repro.py::CudaReproTests::test_unused_cpu_input_cudagraphs, test/inductor/test_cuda_repro.py::CudaReproTests::test_xlnet_lm_stride_repro 2025-07-17T10:03:34.5910850Z 2025-07-17T10:03:34.5911004Z Running inductor/test_cudagraph_trees 1/1 ... [2025-07-17 10:03:34.587927] 2025-07-17T10:03:34.5911298Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:03:34.5911979Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_cudagraph_trees.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:03:34.588287] 2025-07-17T10:08:18.6674960Z 2025-07-17T10:08:18.6679129Z PRINTING LOG FILE of inductor/test_cudagraph_trees 1/1 (test/test-reports/inductor.test_cudagraph_trees_1.1_bba2f2a50707186e_.log) 2025-07-17T10:08:18.6680873Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T10:08:18.6682010Z import pkg_resources 2025-07-17T10:08:18.6682516Z Test results will be stored in test-reports/python-pytest/inductor.test_cudagraph_trees/inductor.test_cudagraph_trees-940a26cd715ceb82.xml 2025-07-17T10:08:18.6682987Z ============================= test session starts ============================== 2025-07-17T10:08:18.6683356Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T10:08:18.6683725Z cachedir: .pytest_cache 2025-07-17T10:08:18.6684149Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T10:08:18.6684581Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T10:08:18.6684783Z configfile: pytest.ini 2025-07-17T10:08:18.6685181Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T10:08:18.6685597Z collecting ... collected 143 items 2025-07-17T10:08:18.6685840Z stepcurrent: Cannot find last run test, not skipping 2025-07-17T10:08:18.6723369Z Running 143 items in this shard: test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_accumulate_grad, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_accumulate_multiple_recordings, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_alias_of_parameter, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_aliased_output_checkpoint, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_aliased_static_parameter, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_aliased_storage_single_weakref, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_aliasing_static_ref, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_amp_cache_disabled, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_backward_gets_cached_cudagraphs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cache_hit_forward_miss_backward, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cached_boxed_forward_device_index, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cached_forward_backward, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_checkpoint_shared_output_storage_deallocation, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_checkpointing_resets_persistent_refs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cleanup, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_compiled_autograd_static_input_params, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_constant_output, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_conv_benchmark, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cpp_wrapper, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cudagraph_capture_sizes, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cudagraph_capture_sizes1, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cudagraph_capture_sizes2, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_dynamic_backward, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_dynamic_warmup, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_empty_cpu_tensor, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_empty_storage, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_end_recording_early, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_error_on_dealloc_use, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_error_on_dealloc_use2, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_execution_into_recording, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_expanded_inputs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_fallback_to_eager_if_recompiling_too_many_times, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_fallback_to_eager_if_recompiling_too_many_times_due_to_cudagraph_managed_tensor, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_fallback_to_eager_if_recompiling_too_many_times_warn_only_once, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_forward_backward, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_forward_backward_not_called_backend_cudagraphs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_forward_backward_not_called_backend_inductor, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_forward_generation, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_forward_with_skipped_cudagraphed_backward, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_frozen_fn, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_function_compiled_multiple_times, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_only, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_op_and_dynamic_shapes, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar1, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar2, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar3, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar4, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar_device_put, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar_mutation, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_tensor_symints, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_custom_op, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_custom_op_dynamoc_shapes, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_custom_op_mutation, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_custom_op_no_split, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_dynamic_shapes, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_forward_backward, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_forward_backward_not_called, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_forward_with_skipped_cudagraphed_backward, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_gc, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_multiple_devices_msg, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_reduce_overhead_mode_effectiveness, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_reorder_cpu_and_gpu, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_reorder_cpu_and_gpu_interleave, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_reorder_custom_op_with_no_dependency, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_reorder_custom_op_with_no_dependency1, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_item, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_nonzero, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_nonzero_backend, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_nonzero_graph_breaks, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_index_put, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_live_outputs_multiple_graphs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_manager_per_device, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mark_step, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_meta_tensor, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_child_node, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_custom_module, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_custom_module_buffer, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_parent_node, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_single_compile_builtin_module, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_single_compile_builtin_module_buffers, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_single_compile_param_inputs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multiple_devices_msg_backend_cudagraphs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multiple_devices_msg_backend_inductor, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multiple_insert_removal_caching, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_backend_cudagraphs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_backend_inductor, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_only_once_backend_cudagraphs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_only_once_backend_inductor, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_backend_cudagraphs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_backend_inductor, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_config_backend_cudagraphs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_config_backend_inductor, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_on_inp_backend_cudagraphs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_on_inp_backend_inductor, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_reinplaced, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_no_rerecord_with_mark_static_address, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_not_fallback_to_eager_if_have_not_recompiling_too_many_times, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_output_alias, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_peristed_output_livenes, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_remove_hooks_on_cached_tensors, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_rerecord_if_static_input_address_changed, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_rng_non_trees, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_rng_trees, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_run_simple, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_separate_recordings, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_side_stream_memory_allocation, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_single_stream_use, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_cpp_wrapper, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_cudagraph_unsafe_ops, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached1, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_symbolic, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_sparsity, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_static_inputs_address_mutation_log, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_storage_access_error, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_tensor_constant_mutation, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_tensor_dies_between_checkpoint, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_tensor_no_longer_in_pool, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_unaligned_static_input_no_cudagraphs, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_unaligned_static_input_non_trees, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_unaligned_static_input_trees, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_unaligned_static_parameter, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_unstable_ptr, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_warmup_stream_sync, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_warn_on_pending_backward, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_warn_once_if_dynamic_shape_limit_reached, test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_workspace_allocation_error, test/inductor/test_cudagraph_trees.py::TestSAC::test_cpu_and_cuda_rng, test/inductor/test_cudagraph_trees.py::TestSAC::test_cudagraph_uneven_forward_backward, test/inductor/test_cudagraph_trees.py::TestSAC::test_cudagraphs_aot_eager_compat_equal, test/inductor/test_cudagraph_trees.py::TestSAC::test_cudagraphs_aot_eager_compat_equal_device_one, test/inductor/test_cudagraph_trees.py::TestSAC::test_graph_partition_cudagraphs_aot_eager_compat_equal, test/inductor/test_cudagraph_trees.py::TestSAC::test_multi_device, test/inductor/test_cudagraph_trees.py::TestSAC::test_retain_graph, test/inductor/test_cudagraph_trees.py::TestSAC::test_simple, test/inductor/test_cudagraph_trees.py::TestSAC::test_uneven_forward_backward_order0, test/inductor/test_cudagraph_trees.py::TestSAC::test_uneven_forward_backward_order1, test/inductor/test_cudagraph_trees.py::TestSAC::test_uneven_forward_backward_order2, test/inductor/test_cudagraph_trees.py::TestSAC::test_uneven_forward_backward_order3, test/inductor/test_cudagraph_trees.py::TestSAC::test_uneven_forward_backward_order4, test/inductor/test_cudagraph_trees.py::TestSAC::test_uneven_forward_backward_order5 2025-07-17T10:08:18.6755670Z 2025-07-17T10:08:18.6755896Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_accumulate_grad PASSED [2.8922s] [ 0%] 2025-07-17T10:08:18.6756560Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_accumulate_multiple_recordings PASSED [4.7397s] [ 1%] 2025-07-17T10:08:18.6757087Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_alias_of_parameter PASSED [0.8443s] [ 2%] 2025-07-17T10:08:18.6757593Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_aliased_output_checkpoint PASSED [0.6474s] [ 2%] 2025-07-17T10:08:18.6758109Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_aliased_static_parameter PASSED [0.4844s] [ 3%] 2025-07-17T10:08:18.6758639Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_aliased_storage_single_weakref PASSED [1.0173s] [ 4%] 2025-07-17T10:08:18.6759146Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_aliasing_static_ref PASSED [2.2369s] [ 4%] 2025-07-17T10:08:18.6759626Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_amp_cache_disabled PASSED [3.8285s] [ 5%] 2025-07-17T10:08:18.6760133Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_backward_gets_cached_cudagraphs PASSED [1.4540s] [ 6%] 2025-07-17T10:08:18.6760675Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cache_hit_forward_miss_backward PASSED [1.9099s] [ 6%] 2025-07-17T10:08:18.6761292Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cached_boxed_forward_device_index SKIPPED [0.0003s] (requires multiple cuda devices) [ 7%] 2025-07-17T10:08:18.6761980Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cached_forward_backward PASSED [2.4801s] [ 8%] 2025-07-17T10:08:18.6762545Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_checkpoint_shared_output_storage_deallocation PASSED [0.6613s] [ 9%] 2025-07-17T10:08:18.6763270Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_checkpointing_resets_persistent_refs SKIPPED [0.1636s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 9%] 2025-07-17T10:08:18.6763902Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cleanup PASSED [1.0658s] [ 10%] 2025-07-17T10:08:18.6764418Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_compiled_autograd_static_input_params PASSED [1.2441s] [ 11%] 2025-07-17T10:08:18.6764939Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_constant_output PASSED [1.0206s] [ 11%] 2025-07-17T10:08:18.6765417Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_conv_benchmark PASSED [1.9618s] [ 12%] 2025-07-17T10:08:18.6765877Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cpp_wrapper PASSED [4.6404s] [ 13%] 2025-07-17T10:08:18.6766358Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cudagraph_capture_sizes PASSED [2.3521s] [ 13%] 2025-07-17T10:08:18.6766864Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cudagraph_capture_sizes1 PASSED [2.3726s] [ 14%] 2025-07-17T10:08:18.6767374Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_cudagraph_capture_sizes2 PASSED [2.3344s] [ 15%] 2025-07-17T10:08:18.6767935Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_dynamic_backward PASSED [2.3638s] [ 16%] 2025-07-17T10:08:18.6769147Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_dynamic_warmup SKIPPED [0.0005s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156693 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 16%] 2025-07-17T10:08:18.6770863Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_empty_cpu_tensor SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156735 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 17%] 2025-07-17T10:08:18.6772560Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_empty_storage SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156755 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 18%] 2025-07-17T10:08:18.6774320Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_end_recording_early SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156778 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 18%] 2025-07-17T10:08:18.6776043Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_error_on_dealloc_use SKIPPED [0.0005s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156801 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 19%] 2025-07-17T10:08:18.6777811Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_error_on_dealloc_use2 SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156808 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 20%] 2025-07-17T10:08:18.6779551Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_execution_into_recording SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156838 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 20%] 2025-07-17T10:08:18.6781272Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_expanded_inputs SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156886 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 21%] 2025-07-17T10:08:18.6783040Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_fallback_to_eager_if_recompiling_too_many_times SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/130749 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 22%] 2025-07-17T10:08:18.6784986Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_fallback_to_eager_if_recompiling_too_many_times_due_to_cudagraph_managed_tensor SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156922 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 23%] 2025-07-17T10:08:18.6786965Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_fallback_to_eager_if_recompiling_too_many_times_warn_only_once SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156954 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 23%] 2025-07-17T10:08:18.6788770Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_forward_backward SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156957 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 24%] 2025-07-17T10:08:18.6790610Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_forward_backward_not_called_backend_cudagraphs SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156984 for platform(s) linux, rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 25%] 2025-07-17T10:08:18.6792441Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_forward_backward_not_called_backend_inductor SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157035 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 25%] 2025-07-17T10:08:18.6794392Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_forward_generation SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157058 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 26%] 2025-07-17T10:08:18.6796166Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_forward_with_skipped_cudagraphed_backward SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157086 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 27%] 2025-07-17T10:08:18.6797354Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_frozen_fn ('RERUN', {'yellow': True}) [0.3105s] [ 27%] 2025-07-17T10:08:18.6797880Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_frozen_fn ('RERUN', {'yellow': True}) [0.3035s] [ 27%] 2025-07-17T10:08:18.6798355Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_frozen_fn FAILED [0.3020s] [ 27%] 2025-07-17T10:08:18.6798615Z 2025-07-17T10:08:18.6798705Z ==================================== RERUNS ==================================== 2025-07-17T10:08:18.6799006Z ______________________ CudaGraphTreeTests.test_frozen_fn _______________________ 2025-07-17T10:08:18.6799293Z Traceback (most recent call last): 2025-07-17T10:08:18.6799657Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 1363, in test_frozen_fn 2025-07-17T10:08:18.6800150Z self.assertTrue(self.get_manager().new_graph_id().id == 1) 2025-07-17T10:08:18.6800481Z ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2025-07-17T10:08:18.6800764Z AttributeError: 'NoneType' object has no attribute 'new_graph_id' 2025-07-17T10:08:18.6800974Z 2025-07-17T10:08:18.6801093Z To execute this test, run the following from the base repo dir: 2025-07-17T10:08:18.6803892Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_cudagraph_trees.py CudaGraphTreeTests.test_frozen_fn 2025-07-17T10:08:18.6804198Z 2025-07-17T10:08:18.6804367Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-07-17T10:08:18.6804719Z ----------------------------- Captured stdout call ----------------------------- 2025-07-17T10:08:18.6804991Z frames [('total', 1), ('ok', 1)] 2025-07-17T10:08:18.6805212Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-07-17T10:08:18.6805538Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-07-17T10:08:18.6806005Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 1), ('extern_calls', 1)] 2025-07-17T10:08:18.6806475Z graph_break [] 2025-07-17T10:08:18.6806645Z aten_mm_info [('aten.mm_10_10_10', 1)] 2025-07-17T10:08:18.6806920Z ______________________ CudaGraphTreeTests.test_frozen_fn _______________________ 2025-07-17T10:08:18.6807197Z Traceback (most recent call last): 2025-07-17T10:08:18.6807612Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 1363, in test_frozen_fn 2025-07-17T10:08:18.6808016Z self.assertTrue(self.get_manager().new_graph_id().id == 1) 2025-07-17T10:08:18.6808275Z ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2025-07-17T10:08:18.6808550Z AttributeError: 'NoneType' object has no attribute 'new_graph_id' 2025-07-17T10:08:18.6808740Z 2025-07-17T10:08:18.6808870Z To execute this test, run the following from the base repo dir: 2025-07-17T10:08:18.6809275Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_cudagraph_trees.py CudaGraphTreeTests.test_frozen_fn 2025-07-17T10:08:18.6809555Z 2025-07-17T10:08:18.6809711Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-07-17T10:08:18.6810050Z ----------------------------- Captured stdout call ----------------------------- 2025-07-17T10:08:18.6810304Z frames [('total', 1), ('ok', 1)] 2025-07-17T10:08:18.6810516Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-07-17T10:08:18.6810845Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-07-17T10:08:18.6811302Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 1), ('extern_calls', 1)] 2025-07-17T10:08:18.6811667Z graph_break [] 2025-07-17T10:08:18.6811831Z aten_mm_info [('aten.mm_10_10_10', 1)] 2025-07-17T10:08:18.6812084Z ----------------------------- Captured stdout call ----------------------------- 2025-07-17T10:08:18.6812339Z frames [('total', 1), ('ok', 1)] 2025-07-17T10:08:18.6812545Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-07-17T10:08:18.6812863Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-07-17T10:08:18.6813329Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 1), ('extern_calls', 1)] 2025-07-17T10:08:18.6813686Z graph_break [] 2025-07-17T10:08:18.6813848Z aten_mm_info [('aten.mm_10_10_10', 1)] 2025-07-17T10:08:18.6814056Z =================================== FAILURES =================================== 2025-07-17T10:08:18.6814347Z ______________________ CudaGraphTreeTests.test_frozen_fn _______________________ 2025-07-17T10:08:18.6814613Z Traceback (most recent call last): 2025-07-17T10:08:18.6814962Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 1363, in test_frozen_fn 2025-07-17T10:08:18.6815364Z self.assertTrue(self.get_manager().new_graph_id().id == 1) 2025-07-17T10:08:18.6815617Z ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2025-07-17T10:08:18.6815968Z AttributeError: 'NoneType' object has no attribute 'new_graph_id' 2025-07-17T10:08:18.6816223Z 2025-07-17T10:08:18.6816338Z To execute this test, run the following from the base repo dir: 2025-07-17T10:08:18.6816731Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_cudagraph_trees.py CudaGraphTreeTests.test_frozen_fn 2025-07-17T10:08:18.6817002Z 2025-07-17T10:08:18.6817151Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-07-17T10:08:18.6817613Z ----------------------------- Captured stdout call ----------------------------- 2025-07-17T10:08:18.6817893Z frames [('total', 1), ('ok', 1)] 2025-07-17T10:08:18.6818116Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-07-17T10:08:18.6818438Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-07-17T10:08:18.6818905Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 1), ('extern_calls', 1)] 2025-07-17T10:08:18.6819272Z graph_break [] 2025-07-17T10:08:18.6819499Z aten_mm_info [('aten.mm_10_10_10', 1)] 2025-07-17T10:08:18.6819765Z ----------------------------- Captured stdout call ----------------------------- 2025-07-17T10:08:18.6820021Z frames [('total', 1), ('ok', 1)] 2025-07-17T10:08:18.6820229Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-07-17T10:08:18.6820537Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-07-17T10:08:18.6821010Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 1), ('extern_calls', 1)] 2025-07-17T10:08:18.6821393Z graph_break [] 2025-07-17T10:08:18.6821570Z aten_mm_info [('aten.mm_10_10_10', 1)] 2025-07-17T10:08:18.6821833Z ----------------------------- Captured stdout call ----------------------------- 2025-07-17T10:08:18.6822100Z frames [('total', 1), ('ok', 1)] 2025-07-17T10:08:18.6822322Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-07-17T10:08:18.6822652Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-07-17T10:08:18.6823125Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 1), ('extern_calls', 1)] 2025-07-17T10:08:18.6823492Z graph_break [] 2025-07-17T10:08:18.6823645Z aten_mm_info [('aten.mm_10_10_10', 1)] 2025-07-17T10:08:18.6824165Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_cudagraph_trees/inductor.test_cudagraph_trees-940a26cd715ceb82.xml - 2025-07-17T10:08:18.6824700Z =========================== short test summary info ============================ 2025-07-17T10:08:18.6825200Z FAILED [0.3020s] inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_frozen_fn - AttributeError: 'NoneType' object has no attribute 'new_graph_id' 2025-07-17T10:08:18.6825705Z 2025-07-17T10:08:18.6825828Z To execute this test, run the following from the base repo dir: 2025-07-17T10:08:18.6826234Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_cudagraph_trees.py CudaGraphTreeTests.test_frozen_fn 2025-07-17T10:08:18.6826530Z 2025-07-17T10:08:18.6826683Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-07-17T10:08:18.6826995Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-07-17T10:08:18.6827279Z ============== 1 failed, 21 passed, 18 skipped, 2 rerun in 43.72s ============== 2025-07-17T10:08:18.6827530Z Got exit code 1 2025-07-17T10:08:18.6827705Z Retrying single test... 2025-07-17T10:08:18.6828546Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T10:08:18.6829367Z import pkg_resources 2025-07-17T10:08:18.6829851Z Test results will be stored in test-reports/python-pytest/inductor.test_cudagraph_trees/inductor.test_cudagraph_trees-95ee880dc791a5ba.xml 2025-07-17T10:08:18.6830372Z ============================= test session starts ============================== 2025-07-17T10:08:18.6830731Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T10:08:18.6831047Z cachedir: .pytest_cache 2025-07-17T10:08:18.6831423Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T10:08:18.6831886Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T10:08:18.6832088Z configfile: pytest.ini 2025-07-17T10:08:18.6832462Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T10:08:18.6832930Z collecting ... collected 143 items / 142 deselected / 1 selected 2025-07-17T10:08:18.6833401Z stepcurrent: skipping 39 already run items. Running only test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_frozen_fn 2025-07-17T10:08:18.6833967Z Running 1 items in this shard 2025-07-17T10:08:18.6834105Z 2025-07-17T10:08:18.6834302Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_frozen_fn PASSED [2.9345s] [100%] 2025-07-17T10:08:18.6834563Z 2025-07-17T10:08:18.6834941Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_cudagraph_trees/inductor.test_cudagraph_trees-95ee880dc791a5ba.xml - 2025-07-17T10:08:18.6835481Z ====================== 1 passed, 142 deselected in 2.96s ======================= 2025-07-17T10:08:18.6835718Z Got exit code 0 2025-07-17T10:08:18.6835956Z Test succeeeded in new process, continuing with the rest of the tests 2025-07-17T10:08:18.6836866Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T10:08:18.6837708Z import pkg_resources 2025-07-17T10:08:18.6838119Z Test results will be stored in test-reports/python-pytest/inductor.test_cudagraph_trees/inductor.test_cudagraph_trees-21c03b048606ddeb.xml 2025-07-17T10:08:18.6838577Z ============================= test session starts ============================== 2025-07-17T10:08:18.6838930Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T10:08:18.6839246Z cachedir: .pytest_cache 2025-07-17T10:08:18.6839622Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T10:08:18.6840027Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T10:08:18.6840222Z configfile: pytest.ini 2025-07-17T10:08:18.6840595Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T10:08:18.6841062Z collecting ... collected 143 items / 40 deselected / 103 selected 2025-07-17T10:08:18.6841349Z stepcurrent: skipping 40 already run items. 2025-07-17T10:08:18.6841560Z Running 103 items in this shard 2025-07-17T10:08:18.6841691Z 2025-07-17T10:08:18.6841943Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_function_compiled_multiple_times PASSED [3.6722s] [ 0%] 2025-07-17T10:08:18.6842645Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition W0717 10:04:49.563000 1492959 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:08:18.6843297Z W0717 10:04:49.565000 1492959 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:08:18.6843627Z PASSED [2.2797s] [ 1%] 2025-07-17T10:08:18.6843957Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_only PASSED [1.3701s] [ 2%] 2025-07-17T10:08:18.6844754Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_op_and_dynamic_shapes W0717 10:04:52.706000 1492959 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:08:18.6845513Z W0717 10:04:52.707000 1492959 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:08:18.6845965Z W0717 10:04:54.200000 1492959 site-packages/torch/_inductor/utils.py:2136] [0/1] DeviceCopy in input program 2025-07-17T10:08:18.6846487Z W0717 10:04:54.202000 1492959 site-packages/torch/_inductor/utils.py:2136] [0/1] DeviceCopy in input program 2025-07-17T10:08:18.6846825Z PASSED [3.4325s] [ 3%] 2025-07-17T10:08:18.6847333Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar1 W0717 10:04:56.069000 1492959 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:08:18.6847878Z PASSED [1.0362s] [ 4%] 2025-07-17T10:08:18.6848385Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar2 W0717 10:04:57.110000 1492959 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:08:18.6849015Z PASSED [1.0044s] [ 5%] 2025-07-17T10:08:18.6850032Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar3 SKIPPED [0.0005s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157339 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 6%] 2025-07-17T10:08:18.6851794Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar4 SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157350 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 7%] 2025-07-17T10:08:18.6853129Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar_device_put W0717 10:04:58.115000 1492959 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:08:18.6853824Z W0717 10:04:58.116000 1492959 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:08:18.6854156Z PASSED [0.2328s] [ 8%] 2025-07-17T10:08:18.6855119Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar_mutation SKIPPED [0.0005s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157359 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 9%] 2025-07-17T10:08:18.6856931Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_cpu_tensor_symints SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157367 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 10%] 2025-07-17T10:08:18.6858701Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_custom_op SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157413 for platform(s) linux, rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 11%] 2025-07-17T10:08:18.6860530Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_custom_op_dynamoc_shapes SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157428 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 12%] 2025-07-17T10:08:18.6862428Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_custom_op_mutation SKIPPED [0.0005s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157449 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 13%] 2025-07-17T10:08:18.6864195Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_custom_op_no_split SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157533 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 14%] 2025-07-17T10:08:18.6866031Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_dynamic_shapes SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157556 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 15%] 2025-07-17T10:08:18.6867769Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_forward_backward SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157616 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 16%] 2025-07-17T10:08:18.6869531Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_forward_backward_not_called SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157643 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 17%] 2025-07-17T10:08:18.6871375Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_forward_with_skipped_cudagraphed_backward SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157723 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 18%] 2025-07-17T10:08:18.6872553Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_gc PASSED [0.9776s] [ 19%] 2025-07-17T10:08:18.6873159Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_multiple_devices_msg SKIPPED [0.0002s] (requires multiple cuda devices) [ 20%] 2025-07-17T10:08:18.6874095Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_reduce_overhead_mode_effectiveness W0717 10:04:59.397000 1492959 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:08:18.6874825Z W0717 10:04:59.398000 1492959 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:08:18.6875159Z PASSED [1.0825s] [ 21%] 2025-07-17T10:08:18.6876114Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_reorder_cpu_and_gpu SKIPPED [0.0005s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157761 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 22%] 2025-07-17T10:08:18.6878563Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_reorder_cpu_and_gpu_interleave SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/152561 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 23%] 2025-07-17T10:08:18.6880555Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_reorder_custom_op_with_no_dependency SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157871 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 24%] 2025-07-17T10:08:18.6882394Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_graph_partition_reorder_custom_op_with_no_dependency1 SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157901 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 25%] 2025-07-17T10:08:18.6883669Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_item PASSED [0.2289s] [ 26%] 2025-07-17T10:08:18.6884235Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_nonzero PASSED [0.2994s] [ 27%] 2025-07-17T10:08:18.6884814Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_nonzero_backend PASSED [0.2244s] [ 28%] 2025-07-17T10:08:18.6885423Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_nonzero_graph_breaks PASSED [1.0826s] [ 29%] 2025-07-17T10:08:18.6885957Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_index_put PASSED [1.3157s] [ 30%] 2025-07-17T10:08:18.6887052Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_live_outputs_multiple_graphs SKIPPED [0.0005s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/158334 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 31%] 2025-07-17T10:08:18.6888235Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_manager_per_device SKIPPED [0.0001s] (requires multiple cuda devices) [ 32%] 2025-07-17T10:08:18.6889367Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mark_step SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/158333 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 33%] 2025-07-17T10:08:18.6890411Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_meta_tensor PASSED [1.0297s] [ 33%] 2025-07-17T10:08:18.6890899Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_child_node PASSED [2.2626s] [ 34%] 2025-07-17T10:08:18.6891421Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_custom_module PASSED [1.7606s] [ 35%] 2025-07-17T10:08:18.6891960Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_custom_module_buffer PASSED [1.9702s] [ 36%] 2025-07-17T10:08:18.6892490Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_parent_node PASSED [2.5314s] [ 37%] 2025-07-17T10:08:18.6893048Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_single_compile_builtin_module PASSED [2.3478s] [ 38%] 2025-07-17T10:08:18.6893660Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_single_compile_builtin_module_buffers PASSED [2.5117s] [ 39%] 2025-07-17T10:08:18.6894385Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multi_dispatch_single_compile_param_inputs PASSED [1.7287s] [ 40%] 2025-07-17T10:08:18.6894962Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multinomial PASSED [0.6344s] [ 41%] 2025-07-17T10:08:18.6895561Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multiple_devices_msg_backend_cudagraphs SKIPPED [0.0002s] (requires multiple cuda devices) [ 42%] 2025-07-17T10:08:18.6896353Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multiple_devices_msg_backend_inductor SKIPPED [0.0001s] (requires multiple cuda devices) [ 43%] 2025-07-17T10:08:18.6896994Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_multiple_insert_removal_caching PASSED [0.1775s] [ 44%] 2025-07-17T10:08:18.6897598Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_backend_cudagraphs PASSED [0.6001s] [ 45%] 2025-07-17T10:08:18.6898263Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_backend_inductor PASSED [1.2073s] [ 46%] 2025-07-17T10:08:18.6899008Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_only_once_backend_cudagraphs PASSED [0.7837s] [ 47%] 2025-07-17T10:08:18.6899702Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_only_once_backend_inductor PASSED [1.1082s] [ 48%] 2025-07-17T10:08:18.6900357Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_backend_cudagraphs PASSED [1.1367s] [ 49%] 2025-07-17T10:08:18.6901005Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_backend_inductor PASSED [1.4685s] [ 50%] 2025-07-17T10:08:18.6901666Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_config_backend_cudagraphs PASSED [0.7770s] [ 51%] 2025-07-17T10:08:18.6902335Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_config_backend_inductor PASSED [1.1163s] [ 52%] 2025-07-17T10:08:18.6902968Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_on_inp_backend_cudagraphs PASSED [0.4324s] [ 53%] 2025-07-17T10:08:18.6903519Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_on_inp_backend_inductor PASSED [1.1158s] [ 54%] 2025-07-17T10:08:18.6904041Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_mutation_reinplaced PASSED [0.7220s] [ 55%] 2025-07-17T10:08:18.6904564Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_no_rerecord_with_mark_static_address PASSED [1.4854s] [ 56%] 2025-07-17T10:08:18.6905167Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_not_fallback_to_eager_if_have_not_recompiling_too_many_times PASSED [1.5299s] [ 57%] 2025-07-17T10:08:18.6905722Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_output_alias PASSED [0.5487s] [ 58%] 2025-07-17T10:08:18.6906209Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_peristed_output_livenes PASSED [0.8401s] [ 59%] 2025-07-17T10:08:18.6906750Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_remove_hooks_on_cached_tensors PASSED [0.7222s] [ 60%] 2025-07-17T10:08:18.6907309Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_rerecord_if_static_input_address_changed PASSED [1.1891s] [ 61%] 2025-07-17T10:08:18.6907839Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_rng_non_trees PASSED [0.2585s] [ 62%] 2025-07-17T10:08:18.6908290Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_rng_trees PASSED [0.2592s] [ 63%] 2025-07-17T10:08:18.6908727Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_run_simple PASSED [1.9610s] [ 64%] 2025-07-17T10:08:18.6909195Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_separate_recordings PASSED [1.9884s] [ 65%] 2025-07-17T10:08:18.6909776Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_side_stream_memory_allocation PASSED [0.5786s] [ 66%] 2025-07-17T10:08:18.6910350Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_single_stream_use PASSED [1.4095s] [ 66%] 2025-07-17T10:08:18.6910831Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_cpp_wrapper PASSED [1.4089s] [ 67%] 2025-07-17T10:08:18.6911329Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_cudagraph_unsafe_ops PASSED [0.5209s] [ 68%] 2025-07-17T10:08:18.6911918Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached1 PASSED [2.2922s] [ 69%] 2025-07-17T10:08:18.6912673Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2 Memory access fault by GPU node-3 (Agent handle: 0x55e69165a600) on address 0x7f5bf2800000. Reason: Unknown. 2025-07-17T10:08:18.6913238Z Fatal Python error: Aborted 2025-07-17T10:08:18.6913357Z 2025-07-17T10:08:18.6913463Z Thread 0x00007f5cd3fff6c0 (most recent call first): 2025-07-17T10:08:18.6913825Z File "/opt/conda/envs/py_3.12/lib/python3.12/concurrent/futures/thread.py", line 90 in _worker 2025-07-17T10:08:18.6914391Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1012 in run 2025-07-17T10:08:18.6914765Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:08:18.6915156Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:08:18.6915383Z 2025-07-17T10:08:18.6915480Z Thread 0x00007f5ce0f3a6c0 (most recent call first): 2025-07-17T10:08:18.6915709Z 2025-07-17T10:08:18.6915810Z 2025-07-17T10:08:18.6915891Z Thread 0x00007f5ce193b6c0 (most recent call first): 2025-07-17T10:08:18.6916223Z File "/opt/conda/envs/py_3.12/lib/python3.12/concurrent/futures/thread.py", line 90 in _worker 2025-07-17T10:08:18.6916619Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1012 in run 2025-07-17T10:08:18.6916992Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:08:18.6917376Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:08:18.6917598Z 2025-07-17T10:08:18.6917683Z Thread 0x00007f5ce243c6c0 (most recent call first): 2025-07-17T10:08:18.6917974Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 359 in wait 2025-07-17T10:08:18.6918329Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 655 in wait 2025-07-17T10:08:18.6918720Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/tqdm/_monitor.py", line 60 in run 2025-07-17T10:08:18.6919137Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:08:18.6919533Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:08:18.6919750Z 2025-07-17T10:08:18.6919854Z Thread 0x00007f5ce2e3d6c0 (most recent call first): 2025-07-17T10:08:18.6920290Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_worker/subproc_pool.py", line 72 in _recv_msg 2025-07-17T10:08:18.6920921Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_worker/subproc_pool.py", line 208 in _read_thread 2025-07-17T10:08:18.6921404Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1012 in run 2025-07-17T10:08:18.6921775Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:08:18.6922156Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:08:18.6922388Z 2025-07-17T10:08:18.6922482Z Thread 0x00007f6ebc819600 (most recent call first): 2025-07-17T10:08:18.6922850Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py", line 1083 in synchronize 2025-07-17T10:08:18.6923396Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2238 in record_function 2025-07-17T10:08:18.6924030Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2182 in _run 2025-07-17T10:08:18.6924615Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2012 in run 2025-07-17T10:08:18.6925105Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/utils.py", line 2797 in run 2025-07-17T10:08:18.6925723Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 388 in deferred_cudagraphify 2025-07-17T10:08:18.6926279Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_fx.py", line 1692 in run 2025-07-17T10:08:18.6926782Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/output_code.py", line 584 in __call__ 2025-07-17T10:08:18.6927371Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 704 in inner_fn 2025-07-17T10:08:18.6927973Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 510 in wrapper 2025-07-17T10:08:18.6928683Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126 in call_func_at_runtime_with_args 2025-07-17T10:08:18.6929316Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 2028 in forward 2025-07-17T10:08:18.6929865Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/autograd/function.py", line 581 in apply 2025-07-17T10:08:18.6930376Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 100 in g 2025-07-17T10:08:18.6930966Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126 in call_func_at_runtime_with_args 2025-07-17T10:08:18.6931614Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 324 in runtime_wrapper 2025-07-17T10:08:18.6932202Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/aot_autograd.py", line 1044 in forward 2025-07-17T10:08:18.6932698Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 1004 in _fn 2025-07-17T10:08:18.6933169Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2613 in forward 2025-07-17T10:08:18.6933655Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784 in _call_impl 2025-07-17T10:08:18.6934186Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773 in _wrapped_call_impl 2025-07-17T10:08:18.6934732Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 803 in compile_wrapper 2025-07-17T10:08:18.6935247Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784 in _call_impl 2025-07-17T10:08:18.6935769Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773 in _wrapped_call_impl 2025-07-17T10:08:18.6936286Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 411 in __call__ 2025-07-17T10:08:18.6936747Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2625 in iter 2025-07-17T10:08:18.6937270Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2630 in test_skip_if_dynamic_shape_limit_reached2 2025-07-17T10:08:18.6937753Z File "/opt/conda/envs/py_3.12/lib/python3.12/contextlib.py", line 81 in inner 2025-07-17T10:08:18.6938200Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3223 in wrapper 2025-07-17T10:08:18.6938731Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 589 in _callTestMethod 2025-07-17T10:08:18.6939215Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 634 in run 2025-07-17T10:08:18.6939738Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3375 in _run_custom 2025-07-17T10:08:18.6940303Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3403 in run 2025-07-17T10:08:18.6940817Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 690 in __call__ 2025-07-17T10:08:18.6941236Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/unittest.py", line 333 in runtest 2025-07-17T10:08:18.6941717Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 169 in pytest_runtest_call 2025-07-17T10:08:18.6942207Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:08:18.6942675Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:08:18.6943200Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:08:18.6943653Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 262 in 2025-07-17T10:08:18.6944102Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 341 in from_call 2025-07-17T10:08:18.6944577Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 261 in call_runtest_hook 2025-07-17T10:08:18.6945063Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 222 in call_and_report 2025-07-17T10:08:18.6945537Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 133 in runtestprotocol 2025-07-17T10:08:18.6946057Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_rerunfailures.py", line 549 in pytest_runtest_protocol 2025-07-17T10:08:18.6946577Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:08:18.6947036Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:08:18.6947487Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:08:18.6947949Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 348 in pytest_runtestloop 2025-07-17T10:08:18.6948417Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:08:18.6948868Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:08:18.6949305Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:08:18.6949730Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 323 in _main 2025-07-17T10:08:18.6950176Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 269 in wrap_session 2025-07-17T10:08:18.6950646Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 316 in pytest_cmdline_main 2025-07-17T10:08:18.6951117Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:08:18.6951568Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:08:18.6952001Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:08:18.6952452Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/config/__init__.py", line 166 in main 2025-07-17T10:08:18.6952974Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 1298 in run_tests 2025-07-17T10:08:18.6953565Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/test_case.py", line 64 in run_tests 2025-07-17T10:08:18.6954274Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/test_case.py", line 15 in run_tests 2025-07-17T10:08:18.6954759Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 4023 in 2025-07-17T10:08:18.6955024Z 2025-07-17T10:08:18.6959626Z Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._bounded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards, torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, markupsafe._speedups, multidict._multidict, yarl._quoting_c, propcache._helpers_c, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket.mask, aiohttp._websocket.reader_c, frozenlist._frozenlist, charset_normalizer.md, requests.packages.charset_normalizer.md, requests.packages.chardet.md, thriftpy2.transport.cybase, thriftpy2.transport.buffered.cybuffered, thriftpy2.transport.framed.cyframed, thriftpy2.transport.memory.cymemory, thriftpy2.transport.sasl.cysasl, thriftpy2.protocol.cybin.cybin, scipy._lib._ccallback_c, yaml._yaml, numba.core.typeconv._typeconv, numba._helperlib, numba._dynfunc, numba._dispatcher, numba.core.runtime._nrt_python, numba.np.ufunc._internal, numba.experimental.jitclass._box, numba.mviewbuf, numba.types.itertools, psutil._psutil_linux, psutil._psutil_posix, hip_utils, __triton_launcher, ciiskcujiyq2itoifpte2pa5zy4dsymaas66sri2c7biik67npyc.kernel, cdoxrolxtunbxqbwoyufqpsenqga2s6mabgahroionzwixele67u.kernel, crjupy7noxddxnsuq4clz2nldoudjncsrk5mzpxvkbdg2uiaoubk.kernel, c7rgkfxlw5vfucdhq732wafiy5rzula33cocleps62kbbqbyxfko.kernel, cm2pm2u5tbcmshmxqz3by67obtz6kxft5z65dmzmrgy7sm5y2hnj.kernel, cwnpm7dwbzvusbyflzh3tkiyodpa6b2ahuwqwwt6nwbofjmva3ng.kernel, cp2wqvlx4oxxp7nmi72ii777sksjewb33mjsqyjcosdwr4wwj7bc.inductor_entry_cpp (total: 64) 2025-07-17T10:08:18.6964451Z Got exit code -6 (SIGIOT) 2025-07-17T10:08:18.6964641Z Retrying single test... 2025-07-17T10:08:18.6965499Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T10:08:18.6966329Z import pkg_resources 2025-07-17T10:08:18.6966740Z Test results will be stored in test-reports/python-pytest/inductor.test_cudagraph_trees/inductor.test_cudagraph_trees-15e87463b2887cf6.xml 2025-07-17T10:08:18.6967222Z ============================= test session starts ============================== 2025-07-17T10:08:18.6967568Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T10:08:18.6967886Z cachedir: .pytest_cache 2025-07-17T10:08:18.6968263Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T10:08:18.6968659Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T10:08:18.6968864Z configfile: pytest.ini 2025-07-17T10:08:18.6969243Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T10:08:18.6969698Z collecting ... collected 143 items / 142 deselected / 1 selected 2025-07-17T10:08:18.6970239Z stepcurrent: skipping 112 already run items. Running only test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2 2025-07-17T10:08:18.6970715Z Running 1 items in this shard 2025-07-17T10:08:18.6970838Z 2025-07-17T10:08:18.6971408Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2 Memory access fault by GPU node-3 (Agent handle: 0x55c9f0d6b7c0) on address 0x7f57a0006000. Reason: Write access to a read-only page. 2025-07-17T10:08:18.6972092Z Fatal Python error: Aborted 2025-07-17T10:08:18.6972210Z 2025-07-17T10:08:18.6972319Z Thread 0x00007f57ba03b6c0 (most recent call first): 2025-07-17T10:08:18.6972546Z 2025-07-17T10:08:18.6972634Z 2025-07-17T10:08:18.6972789Z Thread 0x00007f57baa3c6c0 (most recent call first): 2025-07-17T10:08:18.6973096Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 359 in wait 2025-07-17T10:08:18.6973446Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 655 in wait 2025-07-17T10:08:18.6973831Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/tqdm/_monitor.py", line 60 in run 2025-07-17T10:08:18.6974247Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:08:18.6974639Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:08:18.6974937Z 2025-07-17T10:08:18.6975038Z Thread 0x00007f57bb43d6c0 (most recent call first): 2025-07-17T10:08:18.6975461Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_worker/subproc_pool.py", line 72 in _recv_msg 2025-07-17T10:08:18.6976078Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_worker/subproc_pool.py", line 208 in _read_thread 2025-07-17T10:08:18.6976560Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1012 in run 2025-07-17T10:08:18.6976941Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:08:18.6977334Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:08:18.6977565Z 2025-07-17T10:08:18.6977658Z Thread 0x00007f6993fa0600 (most recent call first): 2025-07-17T10:08:18.6978030Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py", line 1083 in synchronize 2025-07-17T10:08:18.6978572Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2238 in record_function 2025-07-17T10:08:18.6979123Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2182 in _run 2025-07-17T10:08:18.6979643Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2012 in run 2025-07-17T10:08:18.6980133Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/utils.py", line 2797 in run 2025-07-17T10:08:18.6980673Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 388 in deferred_cudagraphify 2025-07-17T10:08:18.6981219Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_fx.py", line 1692 in run 2025-07-17T10:08:18.6981721Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/output_code.py", line 584 in __call__ 2025-07-17T10:08:18.6982286Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 704 in inner_fn 2025-07-17T10:08:18.6982887Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 510 in wrapper 2025-07-17T10:08:18.6983508Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126 in call_func_at_runtime_with_args 2025-07-17T10:08:18.6984137Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 2028 in forward 2025-07-17T10:08:18.6984681Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/autograd/function.py", line 581 in apply 2025-07-17T10:08:18.6985182Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 100 in g 2025-07-17T10:08:18.6985864Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126 in call_func_at_runtime_with_args 2025-07-17T10:08:18.6986588Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 324 in runtime_wrapper 2025-07-17T10:08:18.6987233Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/aot_autograd.py", line 1044 in forward 2025-07-17T10:08:18.6987758Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 1004 in _fn 2025-07-17T10:08:18.6988235Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2613 in forward 2025-07-17T10:08:18.6988719Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784 in _call_impl 2025-07-17T10:08:18.6989256Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773 in _wrapped_call_impl 2025-07-17T10:08:18.6989879Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 803 in compile_wrapper 2025-07-17T10:08:18.6990399Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784 in _call_impl 2025-07-17T10:08:18.6990920Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773 in _wrapped_call_impl 2025-07-17T10:08:18.6991445Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 411 in __call__ 2025-07-17T10:08:18.6991911Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2625 in iter 2025-07-17T10:08:18.6992431Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2630 in test_skip_if_dynamic_shape_limit_reached2 2025-07-17T10:08:18.6992926Z File "/opt/conda/envs/py_3.12/lib/python3.12/contextlib.py", line 81 in inner 2025-07-17T10:08:18.6993393Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3223 in wrapper 2025-07-17T10:08:18.6993965Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 589 in _callTestMethod 2025-07-17T10:08:18.6994380Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 634 in run 2025-07-17T10:08:18.6994856Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3375 in _run_custom 2025-07-17T10:08:18.6995421Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3403 in run 2025-07-17T10:08:18.6996090Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 690 in __call__ 2025-07-17T10:08:18.6996510Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/unittest.py", line 333 in runtest 2025-07-17T10:08:18.6996990Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 169 in pytest_runtest_call 2025-07-17T10:08:18.6997482Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:08:18.6997942Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:08:18.6998398Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:08:18.6998845Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 262 in 2025-07-17T10:08:18.6999286Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 341 in from_call 2025-07-17T10:08:18.6999749Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 261 in call_runtest_hook 2025-07-17T10:08:18.7000236Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 222 in call_and_report 2025-07-17T10:08:18.7000794Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 133 in runtestprotocol 2025-07-17T10:08:18.7001392Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_rerunfailures.py", line 549 in pytest_runtest_protocol 2025-07-17T10:08:18.7001915Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:08:18.7002444Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:08:18.7002897Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:08:18.7003347Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 348 in pytest_runtestloop 2025-07-17T10:08:18.7003813Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:08:18.7004273Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:08:18.7004804Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:08:18.7005234Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 323 in _main 2025-07-17T10:08:18.7005671Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 269 in wrap_session 2025-07-17T10:08:18.7006145Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 316 in pytest_cmdline_main 2025-07-17T10:08:18.7006624Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:08:18.7007110Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:08:18.7007575Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:08:18.7008026Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/config/__init__.py", line 166 in main 2025-07-17T10:08:18.7008556Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 1298 in run_tests 2025-07-17T10:08:18.7009085Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/test_case.py", line 64 in run_tests 2025-07-17T10:08:18.7009584Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/test_case.py", line 15 in run_tests 2025-07-17T10:08:18.7010078Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 4023 in 2025-07-17T10:08:18.7010335Z 2025-07-17T10:08:18.7013592Z Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._bounded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards, torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, markupsafe._speedups, multidict._multidict, yarl._quoting_c, propcache._helpers_c, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket.mask, aiohttp._websocket.reader_c, frozenlist._frozenlist, charset_normalizer.md, requests.packages.charset_normalizer.md, requests.packages.chardet.md, thriftpy2.transport.cybase, thriftpy2.transport.buffered.cybuffered, thriftpy2.transport.framed.cyframed, thriftpy2.transport.memory.cymemory, thriftpy2.transport.sasl.cysasl, thriftpy2.protocol.cybin.cybin, scipy._lib._ccallback_c, yaml._yaml, numba.core.typeconv._typeconv, numba._helperlib, numba._dynfunc, numba._dispatcher, numba.core.runtime._nrt_python, numba.np.ufunc._internal, numba.experimental.jitclass._box, numba.mviewbuf, numba.types.itertools, psutil._psutil_linux, psutil._psutil_posix, hip_utils, __triton_launcher (total: 57) 2025-07-17T10:08:18.7017083Z Got exit code -6 (SIGIOT) 2025-07-17T10:08:18.7017270Z Retrying single test... 2025-07-17T10:08:18.7018165Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T10:08:18.7018995Z import pkg_resources 2025-07-17T10:08:18.7019411Z Test results will be stored in test-reports/python-pytest/inductor.test_cudagraph_trees/inductor.test_cudagraph_trees-59e0c4cb466b5e1d.xml 2025-07-17T10:08:18.7019875Z ============================= test session starts ============================== 2025-07-17T10:08:18.7020225Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T10:08:18.7020540Z cachedir: .pytest_cache 2025-07-17T10:08:18.7021000Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T10:08:18.7021402Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T10:08:18.7021603Z configfile: pytest.ini 2025-07-17T10:08:18.7021982Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T10:08:18.7022437Z collecting ... collected 143 items / 142 deselected / 1 selected 2025-07-17T10:08:18.7022973Z stepcurrent: skipping 112 already run items. Running only test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2 2025-07-17T10:08:18.7023447Z Running 1 items in this shard 2025-07-17T10:08:18.7023564Z 2025-07-17T10:08:18.7024001Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2 Memory access fault by GPU node-3 (Agent handle: 0x561879947c40) on address 0x7f0da5800000. Reason: Unknown. 2025-07-17T10:08:18.7024571Z Fatal Python error: Aborted 2025-07-17T10:08:18.7024691Z 2025-07-17T10:08:18.7024802Z Thread 0x00007f0dbb5fe6c0 (most recent call first): 2025-07-17T10:08:18.7025030Z 2025-07-17T10:08:18.7025124Z 2025-07-17T10:08:18.7025219Z Thread 0x00007f0dbbfff6c0 (most recent call first): 2025-07-17T10:08:18.7025514Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 359 in wait 2025-07-17T10:08:18.7025860Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 655 in wait 2025-07-17T10:08:18.7026246Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/tqdm/_monitor.py", line 60 in run 2025-07-17T10:08:18.7026669Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:08:18.7027057Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:08:18.7027282Z 2025-07-17T10:08:18.7027375Z Thread 0x00007f0dc0cff6c0 (most recent call first): 2025-07-17T10:08:18.7027806Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_worker/subproc_pool.py", line 72 in _recv_msg 2025-07-17T10:08:18.7028410Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_worker/subproc_pool.py", line 208 in _read_thread 2025-07-17T10:08:18.7028891Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1012 in run 2025-07-17T10:08:18.7029260Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:08:18.7029644Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:08:18.7029869Z 2025-07-17T10:08:18.7029955Z Thread 0x00007f1f990a7600 (most recent call first): 2025-07-17T10:08:18.7030330Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py", line 1083 in synchronize 2025-07-17T10:08:18.7030992Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2238 in record_function 2025-07-17T10:08:18.7031610Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2182 in _run 2025-07-17T10:08:18.7032129Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2012 in run 2025-07-17T10:08:18.7032674Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/utils.py", line 2797 in run 2025-07-17T10:08:18.7033227Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 388 in deferred_cudagraphify 2025-07-17T10:08:18.7033775Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_fx.py", line 1692 in run 2025-07-17T10:08:18.7034361Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/output_code.py", line 584 in __call__ 2025-07-17T10:08:18.7034924Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 704 in inner_fn 2025-07-17T10:08:18.7035628Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 510 in wrapper 2025-07-17T10:08:18.7036250Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126 in call_func_at_runtime_with_args 2025-07-17T10:08:18.7036879Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 2028 in forward 2025-07-17T10:08:18.7037420Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/autograd/function.py", line 581 in apply 2025-07-17T10:08:18.7037926Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 100 in g 2025-07-17T10:08:18.7038518Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126 in call_func_at_runtime_with_args 2025-07-17T10:08:18.7039165Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 324 in runtime_wrapper 2025-07-17T10:08:18.7039739Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/aot_autograd.py", line 1044 in forward 2025-07-17T10:08:18.7040222Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 1004 in _fn 2025-07-17T10:08:18.7040695Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2613 in forward 2025-07-17T10:08:18.7041181Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784 in _call_impl 2025-07-17T10:08:18.7041720Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773 in _wrapped_call_impl 2025-07-17T10:08:18.7042276Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 803 in compile_wrapper 2025-07-17T10:08:18.7042782Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784 in _call_impl 2025-07-17T10:08:18.7043308Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773 in _wrapped_call_impl 2025-07-17T10:08:18.7043833Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 411 in __call__ 2025-07-17T10:08:18.7044303Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2625 in iter 2025-07-17T10:08:18.7044827Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2630 in test_skip_if_dynamic_shape_limit_reached2 2025-07-17T10:08:18.7045322Z File "/opt/conda/envs/py_3.12/lib/python3.12/contextlib.py", line 81 in inner 2025-07-17T10:08:18.7045780Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3223 in wrapper 2025-07-17T10:08:18.7046363Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 589 in _callTestMethod 2025-07-17T10:08:18.7046821Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 634 in run 2025-07-17T10:08:18.7047295Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3375 in _run_custom 2025-07-17T10:08:18.7047922Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3403 in run 2025-07-17T10:08:18.7048384Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 690 in __call__ 2025-07-17T10:08:18.7048800Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/unittest.py", line 333 in runtest 2025-07-17T10:08:18.7049278Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 169 in pytest_runtest_call 2025-07-17T10:08:18.7049761Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:08:18.7050286Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:08:18.7050773Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:08:18.7051212Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 262 in 2025-07-17T10:08:18.7051695Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 341 in from_call 2025-07-17T10:08:18.7052161Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 261 in call_runtest_hook 2025-07-17T10:08:18.7052644Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 222 in call_and_report 2025-07-17T10:08:18.7053121Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 133 in runtestprotocol 2025-07-17T10:08:18.7053645Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_rerunfailures.py", line 549 in pytest_runtest_protocol 2025-07-17T10:08:18.7054153Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:08:18.7054616Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:08:18.7055049Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:08:18.7055514Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 348 in pytest_runtestloop 2025-07-17T10:08:18.7055997Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:08:18.7056448Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:08:18.7056887Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:08:18.7057312Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 323 in _main 2025-07-17T10:08:18.7057747Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 269 in wrap_session 2025-07-17T10:08:18.7058207Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 316 in pytest_cmdline_main 2025-07-17T10:08:18.7058668Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:08:18.7059105Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:08:18.7059538Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:08:18.7059987Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/config/__init__.py", line 166 in main 2025-07-17T10:08:18.7060566Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 1298 in run_tests 2025-07-17T10:08:18.7061158Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/test_case.py", line 64 in run_tests 2025-07-17T10:08:18.7061656Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/test_case.py", line 15 in run_tests 2025-07-17T10:08:18.7062127Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 4023 in 2025-07-17T10:08:18.7062377Z 2025-07-17T10:08:18.7065628Z Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._bounded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards, torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, markupsafe._speedups, multidict._multidict, yarl._quoting_c, propcache._helpers_c, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket.mask, aiohttp._websocket.reader_c, frozenlist._frozenlist, charset_normalizer.md, requests.packages.charset_normalizer.md, requests.packages.chardet.md, thriftpy2.transport.cybase, thriftpy2.transport.buffered.cybuffered, thriftpy2.transport.framed.cyframed, thriftpy2.transport.memory.cymemory, thriftpy2.transport.sasl.cysasl, thriftpy2.protocol.cybin.cybin, scipy._lib._ccallback_c, yaml._yaml, numba.core.typeconv._typeconv, numba._helperlib, numba._dynfunc, numba._dispatcher, numba.core.runtime._nrt_python, numba.np.ufunc._internal, numba.experimental.jitclass._box, numba.mviewbuf, numba.types.itertools, psutil._psutil_linux, psutil._psutil_posix, hip_utils, __triton_launcher (total: 57) 2025-07-17T10:08:18.7069068Z Got exit code -6 (SIGIOT) 2025-07-17T10:08:18.7069396Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-07-17T10:08:18.7070388Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T10:08:18.7071201Z import pkg_resources 2025-07-17T10:08:18.7083198Z Test results will be stored in test-reports/python-pytest/inductor.test_cudagraph_trees/inductor.test_cudagraph_trees-c48ea00dfabc81b9.xml 2025-07-17T10:08:18.7083747Z ============================= test session starts ============================== 2025-07-17T10:08:18.7084121Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T10:08:18.7084445Z cachedir: .pytest_cache 2025-07-17T10:08:18.7084830Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T10:08:18.7085248Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T10:08:18.7085456Z configfile: pytest.ini 2025-07-17T10:08:18.7085835Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T10:08:18.7086302Z collecting ... collected 143 items / 113 deselected / 30 selected 2025-07-17T10:08:18.7086578Z stepcurrent: skipping 113 already run items. 2025-07-17T10:08:18.7086803Z Running 30 items in this shard 2025-07-17T10:08:18.7086933Z 2025-07-17T10:08:18.7087150Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_symbolic PASSED [3.0020s] [ 3%] 2025-07-17T10:08:18.7087602Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_sparsity PASSED [0.4162s] [ 6%] 2025-07-17T10:08:18.7088099Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_static_inputs_address_mutation_log PASSED [2.2955s] [ 10%] 2025-07-17T10:08:18.7088767Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_storage_access_error PASSED [0.1594s] [ 13%] 2025-07-17T10:08:18.7089345Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_tensor_constant_mutation PASSED [1.0087s] [ 16%] 2025-07-17T10:08:18.7089871Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_tensor_dies_between_checkpoint PASSED [1.1394s] [ 20%] 2025-07-17T10:08:18.7090463Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_tensor_no_longer_in_pool PASSED [1.3540s] [ 23%] 2025-07-17T10:08:18.7091016Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_unaligned_static_input_no_cudagraphs PASSED [1.2612s] [ 26%] 2025-07-17T10:08:18.7091581Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_unaligned_static_input_non_trees PASSED [0.9848s] [ 30%] 2025-07-17T10:08:18.7092113Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_unaligned_static_input_trees PASSED [1.2756s] [ 33%] 2025-07-17T10:08:18.7092649Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_unaligned_static_parameter PASSED [0.8155s] [ 36%] 2025-07-17T10:08:18.7093220Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_unstable_ptr PASSED [0.9158s] [ 40%] 2025-07-17T10:08:18.7093697Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_warmup_stream_sync PASSED [0.7500s] [ 43%] 2025-07-17T10:08:18.7094190Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_warn_on_pending_backward PASSED [1.0392s] [ 46%] 2025-07-17T10:08:18.7094738Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_warn_once_if_dynamic_shape_limit_reached PASSED [7.4671s] [ 50%] 2025-07-17T10:08:18.7095432Z inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_workspace_allocation_error SKIPPED [0.2672s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 53%] 2025-07-17T10:08:18.7096046Z inductor/test_cudagraph_trees.py::TestSAC::test_cpu_and_cuda_rng PASSED [0.1622s] [ 56%] 2025-07-17T10:08:18.7096503Z inductor/test_cudagraph_trees.py::TestSAC::test_cudagraph_uneven_forward_backward PASSED [0.3219s] [ 60%] 2025-07-17T10:08:18.7096990Z inductor/test_cudagraph_trees.py::TestSAC::test_cudagraphs_aot_eager_compat_equal PASSED [1.8242s] [ 63%] 2025-07-17T10:08:18.7097571Z inductor/test_cudagraph_trees.py::TestSAC::test_cudagraphs_aot_eager_compat_equal_device_one SKIPPED [0.0002s] (requires multiple cuda devices) [ 66%] 2025-07-17T10:08:18.7098194Z inductor/test_cudagraph_trees.py::TestSAC::test_graph_partition_cudagraphs_aot_eager_compat_equal PASSED [1.8265s] [ 70%] 2025-07-17T10:08:18.7098738Z inductor/test_cudagraph_trees.py::TestSAC::test_multi_device SKIPPED [0.0002s] (requires multiple cuda devices) [ 73%] 2025-07-17T10:08:18.7099219Z inductor/test_cudagraph_trees.py::TestSAC::test_retain_graph PASSED [0.0801s] [ 76%] 2025-07-17T10:08:18.7099598Z inductor/test_cudagraph_trees.py::TestSAC::test_simple PASSED [0.2873s] [ 80%] 2025-07-17T10:08:18.7100023Z inductor/test_cudagraph_trees.py::TestSAC::test_uneven_forward_backward_order0 PASSED [0.1048s] [ 83%] 2025-07-17T10:08:18.7100499Z inductor/test_cudagraph_trees.py::TestSAC::test_uneven_forward_backward_order1 PASSED [0.1016s] [ 86%] 2025-07-17T10:08:18.7100955Z inductor/test_cudagraph_trees.py::TestSAC::test_uneven_forward_backward_order2 PASSED [0.1023s] [ 90%] 2025-07-17T10:08:18.7101413Z inductor/test_cudagraph_trees.py::TestSAC::test_uneven_forward_backward_order3 PASSED [0.1016s] [ 93%] 2025-07-17T10:08:18.7101872Z inductor/test_cudagraph_trees.py::TestSAC::test_uneven_forward_backward_order4 PASSED [0.1017s] [ 96%] 2025-07-17T10:08:18.7102321Z inductor/test_cudagraph_trees.py::TestSAC::test_uneven_forward_backward_order5 PASSED [0.1013s] [100%] 2025-07-17T10:08:18.7102590Z 2025-07-17T10:08:18.7102981Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_cudagraph_trees/inductor.test_cudagraph_trees-c48ea00dfabc81b9.xml - 2025-07-17T10:08:18.7103543Z ================ 27 passed, 3 skipped, 113 deselected in 29.34s ================ 2025-07-17T10:08:18.7104112Z The following tests failed and then succeeded when run in a new process['test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_frozen_fn'] 2025-07-17T10:08:18.7104854Z The following tests failed consistently: ['test/inductor/test_cudagraph_trees.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2'] 2025-07-17T10:08:18.7105230Z 2025-07-17T10:08:18.7105597Z FINISHED PRINTING LOG FILE of inductor/test_cudagraph_trees 1/1 (test/test-reports/inductor.test_cudagraph_trees_1.1_bba2f2a50707186e_.log) 2025-07-17T10:08:18.7105954Z 2025-07-17T10:08:18.7106145Z Running inductor/test_cudagraph_trees_expandable_segments 1/1 ... [2025-07-17 10:08:18.668395] 2025-07-17T10:08:18.7106488Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:08:18.7106793Z GITHUB_RUN_ID, GITHUB_RUN_ATTEMPT, or ARTIFACTS_FILE_SUFFIX not set, not uploading 2025-07-17T10:08:18.7107099Z Uploading artifacts took 0.00 seconds 2025-07-17T10:08:18.7107314Z inductor/test_cudagraph_trees 1/1 failed! 2025-07-17T10:08:18.7108060Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_cudagraph_trees_expandable_segments.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:08:18.668704] 2025-07-17T10:10:36.7731488Z 2025-07-17T10:10:36.7732257Z inductor/test_compiled_optimizers 2/3 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_compiled_optimizers_2.3_b51ac31df9436483_.log 2025-07-17T10:10:36.7804187Z Running 232 items in this shard: test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_S429861, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_maximize_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_maximize_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_recompile, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_rho_weight_decay_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_rho_weight_decay_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_tensor_lr_capturable_cuda_cosineannealinglr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_tensor_lr_capturable_cuda_lambdalr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_tensor_lr_capturable_foreach_cuda_cosineannealinglr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_tensor_lr_capturable_foreach_cuda_exponentiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_tensor_lr_capturable_foreach_cuda_onecyclelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_tensor_lr_capturable_foreach_cuda_reducelronplateau, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_weight_decay_capturable_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_weight_decay_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_weight_decay_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adadelta_weight_decay_maximize_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_initial_accumulator_value_weight_decay_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_lr_decay_weight_decay_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_recompile, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_tensor_lr_cpu_exponentiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_tensor_lr_cpu_lambdalr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_tensor_lr_cpu_onecyclelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_tensor_lr_cuda_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_tensor_lr_cuda_polynomiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_tensor_lr_cuda_steplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_tensor_lr_foreach_cuda_constantlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_tensor_lr_foreach_cuda_cosineannealingwarmrestarts, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_tensor_lr_foreach_cuda_multisteplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_tensor_lr_foreach_cuda_polynomiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_weight_decay_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_weight_decay_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adagrad_weight_decay_maximize_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_capturable_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_amsgrad_capturable_cuda_polynomiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_amsgrad_capturable_foreach_cuda_constantlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_amsgrad_capturable_foreach_cuda_lambdalr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_amsgrad_capturable_foreach_cuda_multisteplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_amsgrad_capturable_foreach_cuda_onecyclelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_amsgrad_capturable_foreach_cuda_reducelronplateau, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_amsgrad_capturable_cuda_multiplicativelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_amsgrad_capturable_foreach_cuda_constantlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_amsgrad_capturable_foreach_cuda_cosineannealinglr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_amsgrad_capturable_foreach_cuda_linearlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_capturable_cuda_lambdalr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_capturable_cuda_linearlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_capturable_cuda_multiplicativelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_capturable_cuda_polynomiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_capturable_cuda_steplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_capturable_foreach_cuda_constantlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_capturable_foreach_cuda_cosineannealingwarmrestarts, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_capturable_foreach_cuda_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_capturable_foreach_cuda_lambdalr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_capturable_foreach_cuda_onecyclelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_capturable_foreach_cuda_reducelronplateau, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_tensor_lr_tensor_betas_capturable_foreach_cuda_steplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_weight_decay_amsgrad_capturable_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_weight_decay_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adam_weight_decay_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_capturable_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_recompile, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_tensor_lr_weight_decay_capturable_cuda_exponentiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_tensor_lr_weight_decay_capturable_cuda_lambdalr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_tensor_lr_weight_decay_capturable_cuda_linearlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_tensor_lr_weight_decay_capturable_cuda_onecyclelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_tensor_lr_weight_decay_capturable_foreach_cuda_constantlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_tensor_lr_weight_decay_capturable_foreach_cuda_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_tensor_lr_weight_decay_capturable_foreach_cuda_multiplicativelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_tensor_lr_weight_decay_capturable_foreach_cuda_onecyclelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_tensor_lr_weight_decay_capturable_foreach_cuda_reducelronplateau, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_weight_decay_capturable_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_weight_decay_maximize_capturable_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_weight_decay_maximize_capturable_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamax_weight_decay_maximize_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_capturable_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_recompile, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_amsgrad_capturable_cuda_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_amsgrad_capturable_cuda_onecyclelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_amsgrad_capturable_foreach_cuda_constantlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_amsgrad_capturable_foreach_cuda_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_amsgrad_capturable_foreach_cuda_exponentiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_amsgrad_capturable_foreach_cuda_polynomiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_amsgrad_capturable_foreach_cuda_reducelronplateau, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_amsgrad_capturable_cuda_cosineannealingwarmrestarts, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_amsgrad_capturable_cuda_lambdalr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_amsgrad_capturable_cuda_multiplicativelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_amsgrad_capturable_cuda_multisteplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_amsgrad_capturable_foreach_cuda_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_amsgrad_capturable_foreach_cuda_exponentiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_amsgrad_capturable_foreach_cuda_lambdalr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_capturable_cuda_constantlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_capturable_cuda_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_capturable_cuda_onecyclelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_capturable_cuda_reducelronplateau, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_capturable_cuda_steplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_capturable_foreach_cuda_cosineannealinglr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_tensor_lr_tensor_betas_capturable_foreach_cuda_exponentiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_weight_decay_amsgrad_capturable_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_weight_decay_amsgrad_capturable_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_weight_decay_amsgrad_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_weight_decay_amsgrad_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_weight_decay_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_weight_decay_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_adamw_weight_decay_maximize_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_lambd_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_maximize_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_recompile_default, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_recompile_foreach, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_recompile_single, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_t0_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_tensor_lr_weight_decay_maximize_capturable_cuda_cosineannealingwarmrestarts, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_tensor_lr_weight_decay_maximize_capturable_cuda_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_tensor_lr_weight_decay_maximize_capturable_cuda_multisteplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_tensor_lr_weight_decay_maximize_capturable_cuda_onecyclelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_tensor_lr_weight_decay_maximize_capturable_foreach_cuda_polynomiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_tensor_lr_weight_decay_maximize_capturable_foreach_cuda_steplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_weight_decay_capturable_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_weight_decay_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_weight_decay_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_asgd_weight_decay_maximize_capturable_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_compile_time_smoketest, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_get_value_on_static_address, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_capturable_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_capturable_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_cuda_cosineannealinglr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_cuda_cosineannealingwarmrestarts, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_cuda_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_cuda_exponentiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_cuda_linearlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_cuda_onecyclelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_cuda_polynomiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_cuda_reducelronplateau, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_cuda_steplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_foreach_cuda_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_foreach_cuda_multiplicativelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_foreach_cuda_onecyclelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_foreach_cuda_polynomiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_foreach_cuda_reducelronplateau, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_tensor_lr_weight_decay_momentum_decay_decoupled_weight_decay_capturable_foreach_cuda_steplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_weight_decay_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_weight_decay_maximize_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_weight_decay_maximize_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_weight_decay_momentum_decay_capturable_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_weight_decay_momentum_decay_capturable_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_weight_decay_momentum_decay_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_weight_decay_momentum_decay_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_weight_decay_momentum_decay_decoupled_weight_decay_capturable_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_nadam_weight_decay_momentum_decay_decoupled_weight_decay_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_capturable_weight_decay_decoupled_weight_decay_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_capturable_weight_decay_decoupled_weight_decay_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_eps_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_eps_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_tensor_lr_capturable_weight_decay_decoupled_weight_decay_cuda_constantlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_tensor_lr_capturable_weight_decay_decoupled_weight_decay_cuda_cosineannealinglr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_tensor_lr_capturable_weight_decay_decoupled_weight_decay_cuda_exponentiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_tensor_lr_capturable_weight_decay_decoupled_weight_decay_cuda_multiplicativelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_tensor_lr_capturable_weight_decay_decoupled_weight_decay_cuda_polynomiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_tensor_lr_capturable_weight_decay_decoupled_weight_decay_cuda_steplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_tensor_lr_capturable_weight_decay_decoupled_weight_decay_foreach_cuda_cosineannealinglr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_tensor_lr_capturable_weight_decay_decoupled_weight_decay_foreach_cuda_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_tensor_lr_capturable_weight_decay_decoupled_weight_decay_foreach_cuda_reducelronplateau, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_tensor_lr_capturable_weight_decay_decoupled_weight_decay_foreach_cuda_steplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_weight_decay_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_weight_decay_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_weight_decay_maximize_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_radam_weight_decay_maximize_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_capturable_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_maximize_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_maximize_weight_decay_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_maximize_weight_decay_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_recompile, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_tensor_lr_capturable_cuda_cosineannealinglr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_tensor_lr_capturable_cuda_cosineannealingwarmrestarts, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_tensor_lr_capturable_cuda_lambdalr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_tensor_lr_capturable_cuda_multisteplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_tensor_lr_capturable_cuda_polynomiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_tensor_lr_capturable_foreach_cuda_cosineannealingwarmrestarts, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_tensor_lr_capturable_foreach_cuda_exponentiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_tensor_lr_capturable_foreach_cuda_multiplicativelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_tensor_lr_capturable_foreach_cuda_reducelronplateau, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_weight_decay_centered_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_weight_decay_centered_momentum_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_weight_decay_centered_momentum_maximize_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_weight_decay_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_weight_decay_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rmsprop_weight_decay_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rprop_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rprop_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rprop_etas_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rprop_maximize_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rprop_recompile, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rprop_tensor_lr_capturable_cuda_lambdalr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rprop_tensor_lr_capturable_cuda_steplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rprop_tensor_lr_capturable_foreach_cuda_constantlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rprop_tensor_lr_capturable_foreach_cuda_cosineannealinglr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rprop_tensor_lr_capturable_foreach_cuda_cosineannealingwarmrestarts, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_rprop_tensor_lr_capturable_foreach_cuda_multiplicativelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_momentum_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_momentum_dampening_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_momentum_dampening_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_momentum_nesterov_weight_decay_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_momentum_nesterov_weight_decay_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_tensor_lr_cpu_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_tensor_lr_cpu_lambdalr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_tensor_lr_cuda_constantlr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_tensor_lr_cuda_cosineannealinglr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_tensor_lr_cuda_cycliclr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_tensor_lr_cuda_exponentiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_tensor_lr_cuda_onecyclelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_tensor_lr_cuda_steplr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_tensor_lr_foreach_cuda_exponentiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_tensor_lr_foreach_cuda_multiplicativelr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_tensor_lr_foreach_cuda_polynomiallr, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_weight_decay_cpu, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_sgd_weight_decay_maximize_foreach_cuda, test/inductor/test_compiled_optimizers.py::CompiledOptimizerTests::test_static_address_finalizer, test/inductor/test_compiled_optimizers.py::CompiledOptimizerParityTestsCUDA::test_correctness_ASGD_use_closure_False_cuda_float32, test/inductor/test_compiled_optimizers.py::CompiledOptimizerParityTestsCUDA::test_correctness_Adadelta_use_closure_True_cuda_float32, test/inductor/test_compiled_optimizers.py::CompiledOptimizerParityTestsCUDA::test_correctness_AdamW_use_closure_True_cuda_float32, test/inductor/test_compiled_optimizers.py::CompiledOptimizerParityTestsCUDA::test_correctness_Adam_use_closure_False_cuda_float32, test/inductor/test_compiled_optimizers.py::CompiledOptimizerParityTestsCUDA::test_correctness_LBFGS_use_closure_True_cuda_float32, test/inductor/test_compiled_optimizers.py::CompiledOptimizerParityTestsCUDA::test_correctness_NAdam_use_closure_True_cuda_float32, test/inductor/test_compiled_optimizers.py::CompiledOptimizerParityTestsCUDA::test_correctness_SGD_use_closure_True_cuda_float32 2025-07-17T10:10:36.7872771Z 2025-07-17T10:10:36.7872979Z Running inductor/test_custom_post_grad_passes 1/1 ... [2025-07-17 10:10:36.773398] 2025-07-17T10:10:36.7873290Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:10:36.7874138Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_custom_post_grad_passes.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:10:36.773738] 2025-07-17T10:13:01.7486980Z 2025-07-17T10:13:01.7487770Z inductor/test_custom_post_grad_passes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_custom_post_grad_passes_1.1_762c4cf36c770492_.log 2025-07-17T10:13:01.7489820Z Running 6 items in this shard: test/inductor/test_custom_post_grad_passes.py::TestPostGradCustomPrePostPass::test_custom_backend_pass, test/inductor/test_custom_post_grad_passes.py::TestPostGradCustomPrePostPass::test_custom_joint_pass_post, test/inductor/test_custom_post_grad_passes.py::TestPostGradCustomPrePostPass::test_custom_joint_pass_pre, test/inductor/test_custom_post_grad_passes.py::TestPostGradCustomPrePostPass::test_custom_post_pass, test/inductor/test_custom_post_grad_passes.py::TestPostGradCustomPrePostPass::test_custom_pre_grad_pass, test/inductor/test_custom_post_grad_passes.py::TestPostGradCustomPrePostPass::test_custom_pre_pass 2025-07-17T10:13:01.7491395Z 2025-07-17T10:13:01.7491548Z Running inductor/test_cutlass_evt 1/1 ... [2025-07-17 10:13:01.748611] 2025-07-17T10:13:01.7491825Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:13:01.7492523Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_cutlass_evt.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:13:01.748970] 2025-07-17T10:13:09.1793165Z 2025-07-17T10:13:09.1794265Z inductor/test_cutlass_evt 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cutlass_evt_1.1_96e98ba6d371d4e7_.log 2025-07-17T10:13:09.1796868Z Running 8 items in this shard: test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_evt_argument_codegen, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_evt_argument_codegen_return_accumulator, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_evt_codegen, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_example_tensor_creation, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_py_codegen, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_py_codegen_accumulator_return, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_py_codegen_broadcasting, test/inductor/test_cutlass_evt.py::TestCutlassEVT::test_py_codegen_disjoint_read_indexing 2025-07-17T10:13:09.1798664Z 2025-07-17T10:13:09.1798815Z Running inductor/test_debug_trace 1/1 ... [2025-07-17 10:13:09.179238] 2025-07-17T10:13:09.1799089Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:13:09.1799865Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_debug_trace.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:13:09.179584] 2025-07-17T10:13:31.8924982Z 2025-07-17T10:13:31.8925692Z inductor/test_debug_trace 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_debug_trace_1.1_e60ce6e89d7c8f53_.log 2025-07-17T10:13:31.8926764Z Running 3 items in this shard: test/inductor/test_debug_trace.py::TestDebugTrace::test_debug_multi_tempalte, test/inductor/test_debug_trace.py::TestDebugTrace::test_debug_printer_const, test/inductor/test_debug_trace.py::TestDebugTrace::test_debug_trace 2025-07-17T10:13:31.8927902Z 2025-07-17T10:13:31.8928089Z Running inductor/test_decompose_mem_bound_mm 1/1 ... [2025-07-17 10:13:31.892418] 2025-07-17T10:13:31.8928394Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:13:31.8930537Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_decompose_mem_bound_mm.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:13:31.892750] 2025-07-17T10:13:55.7995206Z 2025-07-17T10:13:55.7998779Z PRINTING LOG FILE of inductor/test_cudagraph_trees_expandable_segments 1/1 (test/test-reports/inductor.test_cudagraph_trees_expandable_segments_1.1_e90ccb86446483a7_.log) 2025-07-17T10:13:55.8000112Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T10:13:55.8001045Z import pkg_resources 2025-07-17T10:13:55.8001442Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to ./.pytorch-disabled-tests.json 2025-07-17T10:13:55.8002172Z Test results will be stored in test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-b0d676235ca91de4.xml 2025-07-17T10:13:55.8002722Z ============================= test session starts ============================== 2025-07-17T10:13:55.8003081Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T10:13:55.8003393Z cachedir: .pytest_cache 2025-07-17T10:13:55.8003766Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T10:13:55.8004200Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T10:13:55.8004412Z configfile: pytest.ini 2025-07-17T10:13:55.8004803Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T10:13:55.8005196Z collecting ... collected 129 items 2025-07-17T10:13:55.8005438Z stepcurrent: Cannot find last run test, not skipping 2025-07-17T10:13:55.8041697Z Running 129 items in this shard: test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_accumulate_grad, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_accumulate_multiple_recordings, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_alias_of_parameter, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_aliased_output_checkpoint, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_aliased_static_parameter, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_aliased_storage_single_weakref, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_aliasing_static_ref, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_amp_cache_disabled, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_backward_gets_cached_cudagraphs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cache_hit_forward_miss_backward, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cached_boxed_forward_device_index, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cached_forward_backward, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_checkpoint_shared_output_storage_deallocation, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_checkpointing_resets_persistent_refs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cleanup, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_compiled_autograd_static_input_params, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_constant_output, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_conv_benchmark, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cpp_wrapper, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cudagraph_capture_sizes, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cudagraph_capture_sizes1, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cudagraph_capture_sizes2, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_dynamic_backward, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_dynamic_warmup, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_empty_cpu_tensor, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_empty_storage, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_end_recording_early, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_error_on_dealloc_use, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_error_on_dealloc_use2, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_execution_into_recording, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_expanded_inputs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_fallback_to_eager_if_recompiling_too_many_times, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_fallback_to_eager_if_recompiling_too_many_times_due_to_cudagraph_managed_tensor, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_fallback_to_eager_if_recompiling_too_many_times_warn_only_once, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_forward_backward, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_forward_backward_not_called_backend_cudagraphs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_forward_backward_not_called_backend_inductor, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_forward_generation, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_forward_with_skipped_cudagraphed_backward, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_frozen_fn, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_function_compiled_multiple_times, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_only, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_op_and_dynamic_shapes, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar1, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar2, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar3, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar4, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar_device_put, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar_mutation, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_tensor_symints, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_custom_op, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_custom_op_dynamoc_shapes, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_custom_op_mutation, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_custom_op_no_split, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_dynamic_shapes, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_forward_backward, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_forward_backward_not_called, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_forward_with_skipped_cudagraphed_backward, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_gc, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_multiple_devices_msg, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_reduce_overhead_mode_effectiveness, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_reorder_cpu_and_gpu, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_reorder_cpu_and_gpu_interleave, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_reorder_custom_op_with_no_dependency, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_reorder_custom_op_with_no_dependency1, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_item, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_nonzero, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_nonzero_backend, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_nonzero_graph_breaks, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_index_put, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_live_outputs_multiple_graphs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_manager_per_device, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mark_step, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_meta_tensor, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_child_node, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_custom_module, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_custom_module_buffer, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_parent_node, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_single_compile_builtin_module, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_single_compile_builtin_module_buffers, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_single_compile_param_inputs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multiple_devices_msg_backend_cudagraphs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multiple_devices_msg_backend_inductor, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multiple_insert_removal_caching, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_backend_cudagraphs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_backend_inductor, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_only_once_backend_cudagraphs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_only_once_backend_inductor, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_backend_cudagraphs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_backend_inductor, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_config_backend_cudagraphs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_config_backend_inductor, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_on_inp_backend_cudagraphs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_on_inp_backend_inductor, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_reinplaced, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_no_rerecord_with_mark_static_address, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_not_fallback_to_eager_if_have_not_recompiling_too_many_times, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_output_alias, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_peristed_output_livenes, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_remove_hooks_on_cached_tensors, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_rerecord_if_static_input_address_changed, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_rng_non_trees, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_rng_trees, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_run_simple, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_separate_recordings, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_side_stream_memory_allocation, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_single_stream_use, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_cpp_wrapper, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_cudagraph_unsafe_ops, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached1, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_symbolic, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_sparsity, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_static_inputs_address_mutation_log, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_storage_access_error, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_tensor_constant_mutation, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_tensor_dies_between_checkpoint, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_tensor_no_longer_in_pool, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_unaligned_static_input_no_cudagraphs, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_unaligned_static_input_non_trees, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_unaligned_static_input_trees, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_unaligned_static_parameter, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_unstable_ptr, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_warmup_stream_sync, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_warn_on_pending_backward, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_warn_once_if_dynamic_shape_limit_reached, test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_workspace_allocation_error 2025-07-17T10:13:55.8077359Z 2025-07-17T10:13:55.8077735Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_accumulate_grad <- test/inductor/test_cudagraph_trees.py PASSED [2.8723s] [ 0%] 2025-07-17T10:13:55.8078522Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_accumulate_multiple_recordings <- test/inductor/test_cudagraph_trees.py PASSED [4.7279s] [ 1%] 2025-07-17T10:13:55.8079321Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_alias_of_parameter <- test/inductor/test_cudagraph_trees.py PASSED [0.8394s] [ 2%] 2025-07-17T10:13:55.8080084Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_aliased_output_checkpoint <- test/inductor/test_cudagraph_trees.py PASSED [0.6352s] [ 3%] 2025-07-17T10:13:55.8080859Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_aliased_static_parameter <- test/inductor/test_cudagraph_trees.py PASSED [0.4740s] [ 3%] 2025-07-17T10:13:55.8081638Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_aliased_storage_single_weakref <- test/inductor/test_cudagraph_trees.py PASSED [1.0107s] [ 4%] 2025-07-17T10:13:55.8082408Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_aliasing_static_ref <- test/inductor/test_cudagraph_trees.py PASSED [2.1720s] [ 5%] 2025-07-17T10:13:55.8083156Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_amp_cache_disabled <- test/inductor/test_cudagraph_trees.py PASSED [3.9024s] [ 6%] 2025-07-17T10:13:55.8083929Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_backward_gets_cached_cudagraphs <- test/inductor/test_cudagraph_trees.py PASSED [1.4189s] [ 6%] 2025-07-17T10:13:55.8084828Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cache_hit_forward_miss_backward <- test/inductor/test_cudagraph_trees.py PASSED [1.8550s] [ 7%] 2025-07-17T10:13:55.8085779Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cached_boxed_forward_device_index <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (requires multiple cuda devices) [ 8%] 2025-07-17T10:13:55.8086710Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cached_forward_backward <- test/inductor/test_cudagraph_trees.py PASSED [2.5157s] [ 9%] 2025-07-17T10:13:55.8087544Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_checkpoint_shared_output_storage_deallocation <- test/inductor/test_cudagraph_trees.py PASSED [0.6669s] [ 10%] 2025-07-17T10:13:55.8088538Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_checkpointing_resets_persistent_refs <- test/inductor/test_cudagraph_trees.py SKIPPED [0.1695s] (skipIfRocm: test doesn't currently work on the ROCm stack) [ 10%] 2025-07-17T10:13:55.8089530Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cleanup <- test/inductor/test_cudagraph_trees.py PASSED [1.0174s] [ 11%] 2025-07-17T10:13:55.8090298Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_compiled_autograd_static_input_params <- test/inductor/test_cudagraph_trees.py PASSED [1.2433s] [ 12%] 2025-07-17T10:13:55.8091082Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_constant_output <- test/inductor/test_cudagraph_trees.py PASSED [1.0101s] [ 13%] 2025-07-17T10:13:55.8091810Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_conv_benchmark <- test/inductor/test_cudagraph_trees.py PASSED [1.1482s] [ 13%] 2025-07-17T10:13:55.8092530Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cpp_wrapper <- test/inductor/test_cudagraph_trees.py PASSED [1.6491s] [ 14%] 2025-07-17T10:13:55.8093270Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cudagraph_capture_sizes <- test/inductor/test_cudagraph_trees.py PASSED [2.3339s] [ 15%] 2025-07-17T10:13:55.8094048Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cudagraph_capture_sizes1 <- test/inductor/test_cudagraph_trees.py PASSED [2.4319s] [ 16%] 2025-07-17T10:13:55.8094813Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_cudagraph_capture_sizes2 <- test/inductor/test_cudagraph_trees.py PASSED [2.3497s] [ 17%] 2025-07-17T10:13:55.8095562Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_dynamic_backward <- test/inductor/test_cudagraph_trees.py PASSED [2.3244s] [ 17%] 2025-07-17T10:13:55.8096980Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_dynamic_warmup <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0005s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156693 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 18%] 2025-07-17T10:13:55.8098967Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_empty_cpu_tensor <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156735 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 19%] 2025-07-17T10:13:55.8100973Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_empty_storage <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156755 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 20%] 2025-07-17T10:13:55.8103038Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_end_recording_early <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156778 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 20%] 2025-07-17T10:13:55.8105007Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_error_on_dealloc_use <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156801 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 21%] 2025-07-17T10:13:55.8107038Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_error_on_dealloc_use2 <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156808 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 22%] 2025-07-17T10:13:55.8109013Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_execution_into_recording <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156838 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 23%] 2025-07-17T10:13:55.8110978Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_expanded_inputs <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156886 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 24%] 2025-07-17T10:13:55.8112988Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_fallback_to_eager_if_recompiling_too_many_times <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/130749 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 24%] 2025-07-17T10:13:55.8115255Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_fallback_to_eager_if_recompiling_too_many_times_due_to_cudagraph_managed_tensor <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156922 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 25%] 2025-07-17T10:13:55.8117513Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_fallback_to_eager_if_recompiling_too_many_times_warn_only_once <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156954 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 26%] 2025-07-17T10:13:55.8119696Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_forward_backward <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156957 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 27%] 2025-07-17T10:13:55.8121730Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_forward_backward_not_called_backend_cudagraphs <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/156984 for platform(s) linux, rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 27%] 2025-07-17T10:13:55.8123874Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_forward_backward_not_called_backend_inductor <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157035 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 28%] 2025-07-17T10:13:55.8125899Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_forward_generation <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157058 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 29%] 2025-07-17T10:13:55.8127931Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_forward_with_skipped_cudagraphed_backward <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157086 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 30%] 2025-07-17T10:13:55.8129364Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_frozen_fn <- test/inductor/test_cudagraph_trees.py ('RERUN', {'yellow': True}) [0.3076s] [ 31%] 2025-07-17T10:13:55.8130144Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_frozen_fn <- test/inductor/test_cudagraph_trees.py ('RERUN', {'yellow': True}) [0.3043s] [ 31%] 2025-07-17T10:13:55.8130982Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_frozen_fn <- test/inductor/test_cudagraph_trees.py FAILED [0.3046s] [ 31%] 2025-07-17T10:13:55.8131367Z 2025-07-17T10:13:55.8131457Z ==================================== RERUNS ==================================== 2025-07-17T10:13:55.8131757Z ______________________ CudaGraphTreeTests.test_frozen_fn _______________________ 2025-07-17T10:13:55.8132030Z Traceback (most recent call last): 2025-07-17T10:13:55.8132378Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 1363, in test_frozen_fn 2025-07-17T10:13:55.8132772Z self.assertTrue(self.get_manager().new_graph_id().id == 1) 2025-07-17T10:13:55.8133021Z ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2025-07-17T10:13:55.8133381Z AttributeError: 'NoneType' object has no attribute 'new_graph_id' 2025-07-17T10:13:55.8133626Z 2025-07-17T10:13:55.8133755Z To execute this test, run the following from the base repo dir: 2025-07-17T10:13:55.8134152Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_cudagraph_trees.py CudaGraphTreeTests.test_frozen_fn 2025-07-17T10:13:55.8134425Z 2025-07-17T10:13:55.8134578Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-07-17T10:13:55.8134961Z ----------------------------- Captured stdout call ----------------------------- 2025-07-17T10:13:55.8135228Z frames [('total', 1), ('ok', 1)] 2025-07-17T10:13:55.8135438Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-07-17T10:13:55.8135759Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-07-17T10:13:55.8136215Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 1), ('extern_calls', 1)] 2025-07-17T10:13:55.8136577Z graph_break [] 2025-07-17T10:13:55.8147665Z aten_mm_info [('aten.mm_10_10_10', 1)] 2025-07-17T10:13:55.8147937Z ______________________ CudaGraphTreeTests.test_frozen_fn _______________________ 2025-07-17T10:13:55.8148201Z Traceback (most recent call last): 2025-07-17T10:13:55.8148545Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 1363, in test_frozen_fn 2025-07-17T10:13:55.8148934Z self.assertTrue(self.get_manager().new_graph_id().id == 1) 2025-07-17T10:13:55.8149180Z ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2025-07-17T10:13:55.8149444Z AttributeError: 'NoneType' object has no attribute 'new_graph_id' 2025-07-17T10:13:55.8149631Z 2025-07-17T10:13:55.8149762Z To execute this test, run the following from the base repo dir: 2025-07-17T10:13:55.8150167Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_cudagraph_trees.py CudaGraphTreeTests.test_frozen_fn 2025-07-17T10:13:55.8150440Z 2025-07-17T10:13:55.8150590Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-07-17T10:13:55.8150918Z ----------------------------- Captured stdout call ----------------------------- 2025-07-17T10:13:55.8151170Z frames [('total', 1), ('ok', 1)] 2025-07-17T10:13:55.8151373Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-07-17T10:13:55.8151671Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-07-17T10:13:55.8152122Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 1), ('extern_calls', 1)] 2025-07-17T10:13:55.8152477Z graph_break [] 2025-07-17T10:13:55.8152641Z aten_mm_info [('aten.mm_10_10_10', 1)] 2025-07-17T10:13:55.8152882Z ----------------------------- Captured stdout call ----------------------------- 2025-07-17T10:13:55.8153120Z frames [('total', 1), ('ok', 1)] 2025-07-17T10:13:55.8153313Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-07-17T10:13:55.8153607Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-07-17T10:13:55.8154125Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 1), ('extern_calls', 1)] 2025-07-17T10:13:55.8154487Z graph_break [] 2025-07-17T10:13:55.8154639Z aten_mm_info [('aten.mm_10_10_10', 1)] 2025-07-17T10:13:55.8154842Z =================================== FAILURES =================================== 2025-07-17T10:13:55.8155118Z ______________________ CudaGraphTreeTests.test_frozen_fn _______________________ 2025-07-17T10:13:55.8155379Z Traceback (most recent call last): 2025-07-17T10:13:55.8155717Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 1363, in test_frozen_fn 2025-07-17T10:13:55.8156119Z self.assertTrue(self.get_manager().new_graph_id().id == 1) 2025-07-17T10:13:55.8156363Z ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2025-07-17T10:13:55.8156620Z AttributeError: 'NoneType' object has no attribute 'new_graph_id' 2025-07-17T10:13:55.8156844Z 2025-07-17T10:13:55.8157125Z To execute this test, run the following from the base repo dir: 2025-07-17T10:13:55.8157601Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_cudagraph_trees.py CudaGraphTreeTests.test_frozen_fn 2025-07-17T10:13:55.8157885Z 2025-07-17T10:13:55.8158024Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-07-17T10:13:55.8158343Z ----------------------------- Captured stdout call ----------------------------- 2025-07-17T10:13:55.8158592Z frames [('total', 1), ('ok', 1)] 2025-07-17T10:13:55.8158859Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-07-17T10:13:55.8159162Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-07-17T10:13:55.8159606Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 1), ('extern_calls', 1)] 2025-07-17T10:13:55.8159955Z graph_break [] 2025-07-17T10:13:55.8160112Z aten_mm_info [('aten.mm_10_10_10', 1)] 2025-07-17T10:13:55.8160349Z ----------------------------- Captured stdout call ----------------------------- 2025-07-17T10:13:55.8160675Z frames [('total', 1), ('ok', 1)] 2025-07-17T10:13:55.8160878Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-07-17T10:13:55.8161192Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-07-17T10:13:55.8161643Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 1), ('extern_calls', 1)] 2025-07-17T10:13:55.8161989Z graph_break [] 2025-07-17T10:13:55.8162149Z aten_mm_info [('aten.mm_10_10_10', 1)] 2025-07-17T10:13:55.8162398Z ----------------------------- Captured stdout call ----------------------------- 2025-07-17T10:13:55.8162650Z frames [('total', 1), ('ok', 1)] 2025-07-17T10:13:55.8162843Z stats [('calls_captured', 1), ('unique_graphs', 1)] 2025-07-17T10:13:55.8163133Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-07-17T10:13:55.8163578Z inductor [('benchmarking.InductorBenchmarker.benchmark_gpu', 2), ('fxgraph_cache_miss', 1), ('extern_calls', 1)] 2025-07-17T10:13:55.8163928Z graph_break [] 2025-07-17T10:13:55.8164086Z aten_mm_info [('aten.mm_10_10_10', 1)] 2025-07-17T10:13:55.8164696Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-b0d676235ca91de4.xml - 2025-07-17T10:13:55.8165314Z =========================== short test summary info ============================ 2025-07-17T10:13:55.8165854Z FAILED [0.3046s] inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_frozen_fn - AttributeError: 'NoneType' object has no attribute 'new_graph_id' 2025-07-17T10:13:55.8166277Z 2025-07-17T10:13:55.8166397Z To execute this test, run the following from the base repo dir: 2025-07-17T10:13:55.8166770Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_cudagraph_trees.py CudaGraphTreeTests.test_frozen_fn 2025-07-17T10:13:55.8167050Z 2025-07-17T10:13:55.8167194Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-07-17T10:13:55.8167508Z !!!!!!!!!!!!!!!!!!!!!!!!!! stopping after 1 failures !!!!!!!!!!!!!!!!!!!!!!!!!!! 2025-07-17T10:13:55.8167781Z ============== 1 failed, 21 passed, 18 skipped, 2 rerun in 39.77s ============== 2025-07-17T10:13:55.8168017Z Got exit code 1 2025-07-17T10:13:55.8168180Z Retrying single test... 2025-07-17T10:13:55.8169011Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T10:13:55.8169837Z import pkg_resources 2025-07-17T10:13:55.8170206Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to ./.pytorch-disabled-tests.json 2025-07-17T10:13:55.8170989Z Test results will be stored in test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-24fd6f7e3f76db27.xml 2025-07-17T10:13:55.8171585Z ============================= test session starts ============================== 2025-07-17T10:13:55.8171932Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T10:13:55.8172235Z cachedir: .pytest_cache 2025-07-17T10:13:55.8172660Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T10:13:55.8173062Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T10:13:55.8173255Z configfile: pytest.ini 2025-07-17T10:13:55.8173630Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T10:13:55.8174077Z collecting ... collected 129 items / 128 deselected / 1 selected 2025-07-17T10:13:55.8174587Z stepcurrent: skipping 39 already run items. Running only test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_frozen_fn 2025-07-17T10:13:55.8175102Z Running 1 items in this shard 2025-07-17T10:13:55.8175217Z 2025-07-17T10:13:55.8175553Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_frozen_fn <- test/inductor/test_cudagraph_trees.py PASSED [2.9823s] [100%] 2025-07-17T10:13:55.8175940Z 2025-07-17T10:13:55.8176419Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-24fd6f7e3f76db27.xml - 2025-07-17T10:13:55.8177063Z ====================== 1 passed, 128 deselected in 3.01s ======================= 2025-07-17T10:13:55.8177291Z Got exit code 0 2025-07-17T10:13:55.8177510Z Test succeeeded in new process, continuing with the rest of the tests 2025-07-17T10:13:55.8178405Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T10:13:55.8179225Z import pkg_resources 2025-07-17T10:13:55.8179588Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to ./.pytorch-disabled-tests.json 2025-07-17T10:13:55.8180293Z Test results will be stored in test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-97dae7bba183e687.xml 2025-07-17T10:13:55.8180826Z ============================= test session starts ============================== 2025-07-17T10:13:55.8181163Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T10:13:55.8181465Z cachedir: .pytest_cache 2025-07-17T10:13:55.8181826Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T10:13:55.8182220Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T10:13:55.8182405Z configfile: pytest.ini 2025-07-17T10:13:55.8182772Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T10:13:55.8183210Z collecting ... collected 129 items / 40 deselected / 89 selected 2025-07-17T10:13:55.8183475Z stepcurrent: skipping 40 already run items. 2025-07-17T10:13:55.8183686Z Running 89 items in this shard 2025-07-17T10:13:55.8183800Z 2025-07-17T10:13:55.8184194Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_function_compiled_multiple_times <- test/inductor/test_cudagraph_trees.py PASSED [4.0275s] [ 1%] 2025-07-17T10:13:55.8185221Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition <- test/inductor/test_cudagraph_trees.py W0717 10:09:30.083000 1586731 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:13:55.8186058Z W0717 10:09:30.084000 1586731 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:13:55.8186387Z PASSED [2.5492s] [ 2%] 2025-07-17T10:13:55.8186844Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_only <- test/inductor/test_cudagraph_trees.py PASSED [1.4524s] [ 3%] 2025-07-17T10:13:55.8187881Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_op_and_dynamic_shapes <- test/inductor/test_cudagraph_trees.py W0717 10:09:33.572000 1586731 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:13:55.8188707Z W0717 10:09:33.574000 1586731 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:13:55.8189154Z W0717 10:09:35.266000 1586731 site-packages/torch/_inductor/utils.py:2136] [0/1] DeviceCopy in input program 2025-07-17T10:13:55.8189686Z W0717 10:09:35.267000 1586731 site-packages/torch/_inductor/utils.py:2136] [0/1] DeviceCopy in input program 2025-07-17T10:13:55.8190009Z PASSED [3.8292s] [ 4%] 2025-07-17T10:13:55.8190648Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar1 <- test/inductor/test_cudagraph_trees.py W0717 10:09:37.331000 1586731 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:13:55.8191307Z PASSED [1.1829s] [ 5%] 2025-07-17T10:13:55.8191930Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar2 <- test/inductor/test_cudagraph_trees.py W0717 10:09:38.520000 1586731 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:13:55.8192581Z PASSED [1.1955s] [ 6%] 2025-07-17T10:13:55.8193693Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar3 <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0006s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157339 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 7%] 2025-07-17T10:13:55.8195801Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar4 <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157350 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 8%] 2025-07-17T10:13:55.8197399Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar_device_put <- test/inductor/test_cudagraph_trees.py W0717 10:09:39.717000 1586731 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:13:55.8198222Z W0717 10:09:39.718000 1586731 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:13:55.8198543Z PASSED [0.3016s] [ 10%] 2025-07-17T10:13:55.8199641Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_scalar_mutation <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0005s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157359 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 11%] 2025-07-17T10:13:55.8201750Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_cpu_tensor_symints <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157367 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 12%] 2025-07-17T10:13:55.8203889Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_custom_op <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157413 for platform(s) linux, rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 13%] 2025-07-17T10:13:55.8205921Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_custom_op_dynamoc_shapes <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157428 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 14%] 2025-07-17T10:13:55.8208032Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_custom_op_mutation <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157449 for platform(s) linux, rocm, slow. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 15%] 2025-07-17T10:13:55.8210035Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_custom_op_no_split <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157533 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 16%] 2025-07-17T10:13:55.8212092Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_dynamic_shapes <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157556 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 17%] 2025-07-17T10:13:55.8214078Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_forward_backward <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157616 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 19%] 2025-07-17T10:13:55.8216092Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_forward_backward_not_called <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157643 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 20%] 2025-07-17T10:13:55.8218217Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_forward_with_skipped_cudagraphed_backward <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157723 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 21%] 2025-07-17T10:13:55.8219775Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_gc <- test/inductor/test_cudagraph_trees.py PASSED [1.2280s] [ 22%] 2025-07-17T10:13:55.8220663Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_multiple_devices_msg <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0002s] (requires multiple cuda devices) [ 23%] 2025-07-17T10:13:55.8221801Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_reduce_overhead_mode_effectiveness <- test/inductor/test_cudagraph_trees.py W0717 10:09:41.323000 1586731 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:13:55.8222737Z W0717 10:09:41.324000 1586731 site-packages/torch/_inductor/utils.py:2136] [0/0] DeviceCopy in input program 2025-07-17T10:13:55.8223060Z PASSED [1.1751s] [ 24%] 2025-07-17T10:13:55.8224139Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_reorder_cpu_and_gpu <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0005s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157761 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 25%] 2025-07-17T10:13:55.8226162Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_reorder_cpu_and_gpu_interleave <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/152561 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 26%] 2025-07-17T10:13:55.8228221Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_reorder_custom_op_with_no_dependency <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157871 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 28%] 2025-07-17T10:13:55.8230281Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_graph_partition_reorder_custom_op_with_no_dependency1 <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0003s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/157901 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 29%] 2025-07-17T10:13:55.8231716Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_item <- test/inductor/test_cudagraph_trees.py PASSED [0.3017s] [ 30%] 2025-07-17T10:13:55.8232536Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_nonzero <- test/inductor/test_cudagraph_trees.py PASSED [0.3655s] [ 31%] 2025-07-17T10:13:55.8233387Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_nonzero_backend <- test/inductor/test_cudagraph_trees.py PASSED [0.2907s] [ 32%] 2025-07-17T10:13:55.8234403Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_incompatible_cudagraph_ops_nonzero_graph_breaks <- test/inductor/test_cudagraph_trees.py PASSED [1.1581s] [ 33%] 2025-07-17T10:13:55.8235265Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_index_put <- test/inductor/test_cudagraph_trees.py PASSED [1.4861s] [ 34%] 2025-07-17T10:13:55.8236689Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_live_outputs_multiple_graphs <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0006s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/158334 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 35%] 2025-07-17T10:13:55.8238129Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_manager_per_device <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0001s] (requires multiple cuda devices) [ 37%] 2025-07-17T10:13:55.8239525Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mark_step <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0004s] (Test is disabled because an issue exists disabling it: https://github.com/pytorch/pytorch/issues/158333 for platform(s) rocm. If you're seeing this on your local machine and would like to enable this test, please make sure CI is not set and you are not using the flag --import-disabled-tests.) [ 38%] 2025-07-17T10:13:55.8240902Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_meta_tensor <- test/inductor/test_cudagraph_trees.py PASSED [1.2292s] [ 39%] 2025-07-17T10:13:55.8241651Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_child_node <- test/inductor/test_cudagraph_trees.py PASSED [2.7497s] [ 40%] 2025-07-17T10:13:55.8242435Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_custom_module <- test/inductor/test_cudagraph_trees.py PASSED [2.1263s] [ 41%] 2025-07-17T10:13:55.8243246Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_custom_module_buffer <- test/inductor/test_cudagraph_trees.py PASSED [2.3530s] [ 42%] 2025-07-17T10:13:55.8244046Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_parent_node <- test/inductor/test_cudagraph_trees.py PASSED [2.9273s] [ 43%] 2025-07-17T10:13:55.8244871Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_single_compile_builtin_module <- test/inductor/test_cudagraph_trees.py PASSED [2.6984s] [ 44%] 2025-07-17T10:13:55.8245750Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_single_compile_builtin_module_buffers <- test/inductor/test_cudagraph_trees.py PASSED [2.8867s] [ 46%] 2025-07-17T10:13:55.8246623Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multi_dispatch_single_compile_param_inputs <- test/inductor/test_cudagraph_trees.py PASSED [2.1177s] [ 47%] 2025-07-17T10:13:55.8247414Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multinomial <- test/inductor/test_cudagraph_trees.py PASSED [0.8586s] [ 48%] 2025-07-17T10:13:55.8248270Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multiple_devices_msg_backend_cudagraphs <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0002s] (requires multiple cuda devices) [ 49%] 2025-07-17T10:13:55.8249259Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multiple_devices_msg_backend_inductor <- test/inductor/test_cudagraph_trees.py SKIPPED [0.0001s] (requires multiple cuda devices) [ 50%] 2025-07-17T10:13:55.8250154Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_multiple_insert_removal_caching <- test/inductor/test_cudagraph_trees.py PASSED [0.2426s] [ 51%] 2025-07-17T10:13:55.8251076Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_backend_cudagraphs <- test/inductor/test_cudagraph_trees.py PASSED [0.8032s] [ 52%] 2025-07-17T10:13:55.8252073Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_backend_inductor <- test/inductor/test_cudagraph_trees.py PASSED [1.3967s] [ 53%] 2025-07-17T10:13:55.8253090Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_only_once_backend_cudagraphs <- test/inductor/test_cudagraph_trees.py PASSED [1.0339s] [ 55%] 2025-07-17T10:13:55.8254060Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensor_warn_only_once_backend_inductor <- test/inductor/test_cudagraph_trees.py PASSED [1.3641s] [ 56%] 2025-07-17T10:13:55.8254998Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_backend_cudagraphs <- test/inductor/test_cudagraph_trees.py PASSED [1.5332s] [ 57%] 2025-07-17T10:13:55.8255967Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_backend_inductor <- test/inductor/test_cudagraph_trees.py PASSED [1.8654s] [ 58%] 2025-07-17T10:13:55.8256894Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_config_backend_cudagraphs <- test/inductor/test_cudagraph_trees.py PASSED [1.0442s] [ 59%] 2025-07-17T10:13:55.8257839Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_cudagraph_managed_tensors_config_backend_inductor <- test/inductor/test_cudagraph_trees.py PASSED [1.3888s] [ 60%] 2025-07-17T10:13:55.8258713Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_on_inp_backend_cudagraphs <- test/inductor/test_cudagraph_trees.py PASSED [0.5696s] [ 61%] 2025-07-17T10:13:55.8259521Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_on_inp_backend_inductor <- test/inductor/test_cudagraph_trees.py PASSED [1.3137s] [ 62%] 2025-07-17T10:13:55.8260310Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_mutation_reinplaced <- test/inductor/test_cudagraph_trees.py PASSED [0.8576s] [ 64%] 2025-07-17T10:13:55.8261101Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_no_rerecord_with_mark_static_address <- test/inductor/test_cudagraph_trees.py PASSED [1.7430s] [ 65%] 2025-07-17T10:13:55.8261961Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_not_fallback_to_eager_if_have_not_recompiling_too_many_times <- test/inductor/test_cudagraph_trees.py PASSED [1.9423s] [ 66%] 2025-07-17T10:13:55.8262771Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_output_alias <- test/inductor/test_cudagraph_trees.py PASSED [0.7454s] [ 67%] 2025-07-17T10:13:55.8263514Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_peristed_output_livenes <- test/inductor/test_cudagraph_trees.py PASSED [1.0383s] [ 68%] 2025-07-17T10:13:55.8264301Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_remove_hooks_on_cached_tensors <- test/inductor/test_cudagraph_trees.py PASSED [0.8614s] [ 69%] 2025-07-17T10:13:55.8265117Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_rerecord_if_static_input_address_changed <- test/inductor/test_cudagraph_trees.py PASSED [1.4703s] [ 70%] 2025-07-17T10:13:55.8265895Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_rng_non_trees <- test/inductor/test_cudagraph_trees.py PASSED [0.3262s] [ 71%] 2025-07-17T10:13:55.8266601Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_rng_trees <- test/inductor/test_cudagraph_trees.py PASSED [0.3252s] [ 73%] 2025-07-17T10:13:55.8267363Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_run_simple <- test/inductor/test_cudagraph_trees.py PASSED [2.2097s] [ 74%] 2025-07-17T10:13:55.8268150Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_separate_recordings <- test/inductor/test_cudagraph_trees.py PASSED [2.2134s] [ 75%] 2025-07-17T10:13:55.8268930Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_side_stream_memory_allocation <- test/inductor/test_cudagraph_trees.py PASSED [0.7583s] [ 76%] 2025-07-17T10:13:55.8269764Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_single_stream_use <- test/inductor/test_cudagraph_trees.py PASSED [1.6560s] [ 77%] 2025-07-17T10:13:55.8270506Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_cpp_wrapper <- test/inductor/test_cudagraph_trees.py PASSED [1.4841s] [ 78%] 2025-07-17T10:13:55.8271259Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_cudagraph_unsafe_ops <- test/inductor/test_cudagraph_trees.py PASSED [0.6685s] [ 79%] 2025-07-17T10:13:55.8272119Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached1 <- test/inductor/test_cudagraph_trees.py PASSED [2.6522s] [ 80%] 2025-07-17T10:13:55.8273188Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2 <- test/inductor/test_cudagraph_trees.py Memory access fault by GPU node-3 (Agent handle: 0x557e5e3c3380) on address 0x7fbf7d600000. Reason: Write access to a read-only page. 2025-07-17T10:13:55.8274016Z Fatal Python error: Aborted 2025-07-17T10:13:55.8274140Z 2025-07-17T10:13:55.8274239Z Thread 0x00007fc05f5fe6c0 (most recent call first): 2025-07-17T10:13:55.8274594Z File "/opt/conda/envs/py_3.12/lib/python3.12/concurrent/futures/thread.py", line 90 in _worker 2025-07-17T10:13:55.8274981Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1012 in run 2025-07-17T10:13:55.8275354Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:13:55.8275742Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:13:55.8275963Z 2025-07-17T10:13:55.8276056Z Thread 0x00007fc05ffff6c0 (most recent call first): 2025-07-17T10:13:55.8276270Z 2025-07-17T10:13:55.8276364Z 2025-07-17T10:13:55.8276448Z Thread 0x00007fc06cb3b6c0 (most recent call first): 2025-07-17T10:13:55.8276781Z File "/opt/conda/envs/py_3.12/lib/python3.12/concurrent/futures/thread.py", line 90 in _worker 2025-07-17T10:13:55.8277157Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1012 in run 2025-07-17T10:13:55.8277516Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:13:55.8277890Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:13:55.8278100Z 2025-07-17T10:13:55.8278192Z Thread 0x00007fc06d63c6c0 (most recent call first): 2025-07-17T10:13:55.8278477Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 359 in wait 2025-07-17T10:13:55.8278806Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 655 in wait 2025-07-17T10:13:55.8279174Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/tqdm/_monitor.py", line 60 in run 2025-07-17T10:13:55.8279570Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:13:55.8279938Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:13:55.8280153Z 2025-07-17T10:13:55.8280237Z Thread 0x00007fc06e03d6c0 (most recent call first): 2025-07-17T10:13:55.8280652Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_worker/subproc_pool.py", line 72 in _recv_msg 2025-07-17T10:13:55.8281247Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_worker/subproc_pool.py", line 208 in _read_thread 2025-07-17T10:13:55.8281793Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1012 in run 2025-07-17T10:13:55.8282217Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:13:55.8282585Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:13:55.8282800Z 2025-07-17T10:13:55.8282884Z Thread 0x00007fd2478f7600 (most recent call first): 2025-07-17T10:13:55.8283301Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py", line 1083 in synchronize 2025-07-17T10:13:55.8283843Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2238 in record_function 2025-07-17T10:13:55.8284385Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2182 in _run 2025-07-17T10:13:55.8284893Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2012 in run 2025-07-17T10:13:55.8285446Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/utils.py", line 2797 in run 2025-07-17T10:13:55.8285976Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 388 in deferred_cudagraphify 2025-07-17T10:13:55.8286512Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_fx.py", line 1692 in run 2025-07-17T10:13:55.8287010Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/output_code.py", line 584 in __call__ 2025-07-17T10:13:55.8287591Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 704 in inner_fn 2025-07-17T10:13:55.8288182Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 510 in wrapper 2025-07-17T10:13:55.8288792Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126 in call_func_at_runtime_with_args 2025-07-17T10:13:55.8289409Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 2028 in forward 2025-07-17T10:13:55.8289946Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/autograd/function.py", line 581 in apply 2025-07-17T10:13:55.8290448Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 100 in g 2025-07-17T10:13:55.8291028Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126 in call_func_at_runtime_with_args 2025-07-17T10:13:55.8291663Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 324 in runtime_wrapper 2025-07-17T10:13:55.8292228Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/aot_autograd.py", line 1044 in forward 2025-07-17T10:13:55.8292707Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 1004 in _fn 2025-07-17T10:13:55.8293170Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2613 in forward 2025-07-17T10:13:55.8293641Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784 in _call_impl 2025-07-17T10:13:55.8294164Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773 in _wrapped_call_impl 2025-07-17T10:13:55.8294691Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 803 in compile_wrapper 2025-07-17T10:13:55.8295198Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784 in _call_impl 2025-07-17T10:13:55.8295710Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773 in _wrapped_call_impl 2025-07-17T10:13:55.8296290Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 411 in __call__ 2025-07-17T10:13:55.8296807Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2625 in iter 2025-07-17T10:13:55.8297319Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2630 in test_skip_if_dynamic_shape_limit_reached2 2025-07-17T10:13:55.8297793Z File "/opt/conda/envs/py_3.12/lib/python3.12/contextlib.py", line 81 in inner 2025-07-17T10:13:55.8298294Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3223 in wrapper 2025-07-17T10:13:55.8298779Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 589 in _callTestMethod 2025-07-17T10:13:55.8299156Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 634 in run 2025-07-17T10:13:55.8299609Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3375 in _run_custom 2025-07-17T10:13:55.8300218Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3403 in run 2025-07-17T10:13:55.8300664Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 690 in __call__ 2025-07-17T10:13:55.8301074Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/unittest.py", line 333 in runtest 2025-07-17T10:13:55.8301546Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 169 in pytest_runtest_call 2025-07-17T10:13:55.8302020Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:13:55.8302470Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:13:55.8302908Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:13:55.8303339Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 262 in 2025-07-17T10:13:55.8303777Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 341 in from_call 2025-07-17T10:13:55.8304235Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 261 in call_runtest_hook 2025-07-17T10:13:55.8304709Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 222 in call_and_report 2025-07-17T10:13:55.8305181Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 133 in runtestprotocol 2025-07-17T10:13:55.8305693Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_rerunfailures.py", line 549 in pytest_runtest_protocol 2025-07-17T10:13:55.8306189Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:13:55.8306626Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:13:55.8307066Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:13:55.8307523Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 348 in pytest_runtestloop 2025-07-17T10:13:55.8307984Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:13:55.8308423Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:13:55.8308850Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:13:55.8309267Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 323 in _main 2025-07-17T10:13:55.8309698Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 269 in wrap_session 2025-07-17T10:13:55.8310158Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 316 in pytest_cmdline_main 2025-07-17T10:13:55.8310712Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:13:55.8311226Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:13:55.8311653Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:13:55.8312146Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/config/__init__.py", line 166 in main 2025-07-17T10:13:55.8312666Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 1298 in run_tests 2025-07-17T10:13:55.8313227Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees_expandable_segments.py", line 40 in 2025-07-17T10:13:55.8313522Z 2025-07-17T10:13:55.8318059Z Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._bounded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards, torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, multidict._multidict, yarl._quoting_c, propcache._helpers_c, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket.mask, aiohttp._websocket.reader_c, frozenlist._frozenlist, charset_normalizer.md, requests.packages.charset_normalizer.md, requests.packages.chardet.md, thriftpy2.transport.cybase, thriftpy2.transport.buffered.cybuffered, thriftpy2.transport.framed.cyframed, thriftpy2.transport.memory.cymemory, thriftpy2.transport.sasl.cysasl, thriftpy2.protocol.cybin.cybin, scipy._lib._ccallback_c, yaml._yaml, numba.core.typeconv._typeconv, numba._helperlib, numba._dynfunc, numba._dispatcher, numba.core.runtime._nrt_python, numba.np.ufunc._internal, numba.experimental.jitclass._box, numba.mviewbuf, numba.types.itertools, markupsafe._speedups, psutil._psutil_linux, psutil._psutil_posix, hip_utils, __triton_launcher, ciiskcujiyq2itoifpte2pa5zy4dsymaas66sri2c7biik67npyc.kernel, cdoxrolxtunbxqbwoyufqpsenqga2s6mabgahroionzwixele67u.kernel, crjupy7noxddxnsuq4clz2nldoudjncsrk5mzpxvkbdg2uiaoubk.kernel, c7rgkfxlw5vfucdhq732wafiy5rzula33cocleps62kbbqbyxfko.kernel, cm2pm2u5tbcmshmxqz3by67obtz6kxft5z65dmzmrgy7sm5y2hnj.kernel, cwnpm7dwbzvusbyflzh3tkiyodpa6b2ahuwqwwt6nwbofjmva3ng.kernel, cycswls2lj673bwyjsc5hgszhfv6damfk34iv37f5b7qtt7evb6n.inductor_entry_cpp (total: 64) 2025-07-17T10:13:55.8322756Z Got exit code -6 (SIGIOT) 2025-07-17T10:13:55.8322936Z Retrying single test... 2025-07-17T10:13:55.8323756Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T10:13:55.8324571Z import pkg_resources 2025-07-17T10:13:55.8324950Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to ./.pytorch-disabled-tests.json 2025-07-17T10:13:55.8325671Z Test results will be stored in test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-18154e3dccc2a4e5.xml 2025-07-17T10:13:55.8326222Z ============================= test session starts ============================== 2025-07-17T10:13:55.8326569Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T10:13:55.8326876Z cachedir: .pytest_cache 2025-07-17T10:13:55.8327243Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T10:13:55.8327703Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T10:13:55.8327959Z configfile: pytest.ini 2025-07-17T10:13:55.8328346Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T10:13:55.8328797Z collecting ... collected 129 items / 128 deselected / 1 selected 2025-07-17T10:13:55.8329438Z stepcurrent: skipping 112 already run items. Running only test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2 2025-07-17T10:13:55.8329965Z Running 1 items in this shard 2025-07-17T10:13:55.8330084Z 2025-07-17T10:13:55.8330721Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2 <- test/inductor/test_cudagraph_trees.py Memory access fault by GPU node-3 (Agent handle: 0x55a97ebabc60) on address 0x7f8c31a02000. Reason: Write access to a read-only page. 2025-07-17T10:13:55.8331476Z Fatal Python error: Aborted 2025-07-17T10:13:55.8331655Z 2025-07-17T10:13:55.8331763Z Thread 0x00007f8c475fe6c0 (most recent call first): 2025-07-17T10:13:55.8331987Z 2025-07-17T10:13:55.8332078Z 2025-07-17T10:13:55.8332177Z Thread 0x00007f8c47fff6c0 (most recent call first): 2025-07-17T10:13:55.8332494Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 359 in wait 2025-07-17T10:13:55.8332846Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 655 in wait 2025-07-17T10:13:55.8333236Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/tqdm/_monitor.py", line 60 in run 2025-07-17T10:13:55.8333645Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:13:55.8334036Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:13:55.8334264Z 2025-07-17T10:13:55.8334357Z Thread 0x00007f8c4cdff6c0 (most recent call first): 2025-07-17T10:13:55.8334794Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_worker/subproc_pool.py", line 72 in _recv_msg 2025-07-17T10:13:55.8335404Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_worker/subproc_pool.py", line 208 in _read_thread 2025-07-17T10:13:55.8335876Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1012 in run 2025-07-17T10:13:55.8336245Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:13:55.8336628Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:13:55.8336847Z 2025-07-17T10:13:55.8336934Z Thread 0x00007f9e25963600 (most recent call first): 2025-07-17T10:13:55.8337302Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py", line 1083 in synchronize 2025-07-17T10:13:55.8337931Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2238 in record_function 2025-07-17T10:13:55.8338592Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2182 in _run 2025-07-17T10:13:55.8339347Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2012 in run 2025-07-17T10:13:55.8339945Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/utils.py", line 2797 in run 2025-07-17T10:13:55.8351728Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 388 in deferred_cudagraphify 2025-07-17T10:13:55.8352320Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_fx.py", line 1692 in run 2025-07-17T10:13:55.8352829Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/output_code.py", line 584 in __call__ 2025-07-17T10:13:55.8353384Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 704 in inner_fn 2025-07-17T10:13:55.8354196Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 510 in wrapper 2025-07-17T10:13:55.8354888Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126 in call_func_at_runtime_with_args 2025-07-17T10:13:55.8355576Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 2028 in forward 2025-07-17T10:13:55.8356114Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/autograd/function.py", line 581 in apply 2025-07-17T10:13:55.8356611Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 100 in g 2025-07-17T10:13:55.8357182Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126 in call_func_at_runtime_with_args 2025-07-17T10:13:55.8357816Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 324 in runtime_wrapper 2025-07-17T10:13:55.8358465Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/aot_autograd.py", line 1044 in forward 2025-07-17T10:13:55.8358954Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 1004 in _fn 2025-07-17T10:13:55.8359418Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2613 in forward 2025-07-17T10:13:55.8359898Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784 in _call_impl 2025-07-17T10:13:55.8360425Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773 in _wrapped_call_impl 2025-07-17T10:13:55.8360961Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 803 in compile_wrapper 2025-07-17T10:13:55.8361469Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784 in _call_impl 2025-07-17T10:13:55.8361985Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773 in _wrapped_call_impl 2025-07-17T10:13:55.8362491Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 411 in __call__ 2025-07-17T10:13:55.8362954Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2625 in iter 2025-07-17T10:13:55.8363467Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2630 in test_skip_if_dynamic_shape_limit_reached2 2025-07-17T10:13:55.8363938Z File "/opt/conda/envs/py_3.12/lib/python3.12/contextlib.py", line 81 in inner 2025-07-17T10:13:55.8364388Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3223 in wrapper 2025-07-17T10:13:55.8364869Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 589 in _callTestMethod 2025-07-17T10:13:55.8365250Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 634 in run 2025-07-17T10:13:55.8365714Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3375 in _run_custom 2025-07-17T10:13:55.8366266Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3403 in run 2025-07-17T10:13:55.8366717Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 690 in __call__ 2025-07-17T10:13:55.8367129Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/unittest.py", line 333 in runtest 2025-07-17T10:13:55.8367594Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 169 in pytest_runtest_call 2025-07-17T10:13:55.8368069Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:13:55.8368588Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:13:55.8369087Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:13:55.8369520Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 262 in 2025-07-17T10:13:55.8369950Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 341 in from_call 2025-07-17T10:13:55.8370468Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 261 in call_runtest_hook 2025-07-17T10:13:55.8370956Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 222 in call_and_report 2025-07-17T10:13:55.8371429Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 133 in runtestprotocol 2025-07-17T10:13:55.8371941Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_rerunfailures.py", line 549 in pytest_runtest_protocol 2025-07-17T10:13:55.8372503Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:13:55.8372951Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:13:55.8373389Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:13:55.8373842Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 348 in pytest_runtestloop 2025-07-17T10:13:55.8374298Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:13:55.8374741Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:13:55.8375166Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:13:55.8375585Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 323 in _main 2025-07-17T10:13:55.8376028Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 269 in wrap_session 2025-07-17T10:13:55.8376483Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 316 in pytest_cmdline_main 2025-07-17T10:13:55.8376942Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:13:55.8377378Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:13:55.8377801Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:13:55.8378242Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/config/__init__.py", line 166 in main 2025-07-17T10:13:55.8378754Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 1298 in run_tests 2025-07-17T10:13:55.8379313Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees_expandable_segments.py", line 40 in 2025-07-17T10:13:55.8379614Z 2025-07-17T10:13:55.8382895Z Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._bounded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards, torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, multidict._multidict, yarl._quoting_c, propcache._helpers_c, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket.mask, aiohttp._websocket.reader_c, frozenlist._frozenlist, charset_normalizer.md, requests.packages.charset_normalizer.md, requests.packages.chardet.md, thriftpy2.transport.cybase, thriftpy2.transport.buffered.cybuffered, thriftpy2.transport.framed.cyframed, thriftpy2.transport.memory.cymemory, thriftpy2.transport.sasl.cysasl, thriftpy2.protocol.cybin.cybin, scipy._lib._ccallback_c, yaml._yaml, numba.core.typeconv._typeconv, numba._helperlib, numba._dynfunc, numba._dispatcher, numba.core.runtime._nrt_python, numba.np.ufunc._internal, numba.experimental.jitclass._box, numba.mviewbuf, numba.types.itertools, markupsafe._speedups, psutil._psutil_linux, psutil._psutil_posix, hip_utils, __triton_launcher (total: 57) 2025-07-17T10:13:55.8386266Z Got exit code -6 (SIGIOT) 2025-07-17T10:13:55.8386451Z Retrying single test... 2025-07-17T10:13:55.8387272Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T10:13:55.8388162Z import pkg_resources 2025-07-17T10:13:55.8388539Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to ./.pytorch-disabled-tests.json 2025-07-17T10:13:55.8389251Z Test results will be stored in test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-9b5b93380f101264.xml 2025-07-17T10:13:55.8389795Z ============================= test session starts ============================== 2025-07-17T10:13:55.8390144Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T10:13:55.8390448Z cachedir: .pytest_cache 2025-07-17T10:13:55.8390818Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T10:13:55.8391208Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T10:13:55.8391397Z configfile: pytest.ini 2025-07-17T10:13:55.8391774Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T10:13:55.8392226Z collecting ... collected 129 items / 128 deselected / 1 selected 2025-07-17T10:13:55.8392805Z stepcurrent: skipping 112 already run items. Running only test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2 2025-07-17T10:13:55.8393317Z Running 1 items in this shard 2025-07-17T10:13:55.8393434Z 2025-07-17T10:13:55.8394119Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2 <- test/inductor/test_cudagraph_trees.py Memory access fault by GPU node-3 (Agent handle: 0x5596884bdd90) on address 0x7f29b1a00000. Reason: Write access to a read-only page. 2025-07-17T10:13:55.8394862Z Fatal Python error: Aborted 2025-07-17T10:13:55.8394976Z 2025-07-17T10:13:55.8395071Z Thread 0x00007f29c75fe6c0 (most recent call first): 2025-07-17T10:13:55.8395282Z 2025-07-17T10:13:55.8395376Z 2025-07-17T10:13:55.8395469Z Thread 0x00007f29c7fff6c0 (most recent call first): 2025-07-17T10:13:55.8395757Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 359 in wait 2025-07-17T10:13:55.8396091Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 655 in wait 2025-07-17T10:13:55.8396459Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/tqdm/_monitor.py", line 60 in run 2025-07-17T10:13:55.8396875Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:13:55.8397253Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:13:55.8397476Z 2025-07-17T10:13:55.8397561Z Thread 0x00007f29ccdff6c0 (most recent call first): 2025-07-17T10:13:55.8397971Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_worker/subproc_pool.py", line 72 in _recv_msg 2025-07-17T10:13:55.8398639Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_worker/subproc_pool.py", line 208 in _read_thread 2025-07-17T10:13:55.8399178Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1012 in run 2025-07-17T10:13:55.8399548Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1075 in _bootstrap_inner 2025-07-17T10:13:55.8399934Z File "/opt/conda/envs/py_3.12/lib/python3.12/threading.py", line 1032 in _bootstrap 2025-07-17T10:13:55.8400219Z 2025-07-17T10:13:55.8400308Z Thread 0x00007f3ba5aa2600 (most recent call first): 2025-07-17T10:13:55.8400671Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/cuda/__init__.py", line 1083 in synchronize 2025-07-17T10:13:55.8401198Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2238 in record_function 2025-07-17T10:13:55.8401735Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2182 in _run 2025-07-17T10:13:55.8402312Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 2012 in run 2025-07-17T10:13:55.8402796Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/utils.py", line 2797 in run 2025-07-17T10:13:55.8403322Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/cudagraph_trees.py", line 388 in deferred_cudagraphify 2025-07-17T10:13:55.8403865Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/compile_fx.py", line 1692 in run 2025-07-17T10:13:55.8404350Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_inductor/output_code.py", line 584 in __call__ 2025-07-17T10:13:55.8404895Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 704 in inner_fn 2025-07-17T10:13:55.8405479Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 510 in wrapper 2025-07-17T10:13:55.8406092Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126 in call_func_at_runtime_with_args 2025-07-17T10:13:55.8406710Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 2028 in forward 2025-07-17T10:13:55.8407237Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/autograd/function.py", line 581 in apply 2025-07-17T10:13:55.8407729Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 100 in g 2025-07-17T10:13:55.8408297Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126 in call_func_at_runtime_with_args 2025-07-17T10:13:55.8408918Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 324 in runtime_wrapper 2025-07-17T10:13:55.8409480Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_functorch/aot_autograd.py", line 1044 in forward 2025-07-17T10:13:55.8409959Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 1004 in _fn 2025-07-17T10:13:55.8410408Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2613 in forward 2025-07-17T10:13:55.8410879Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784 in _call_impl 2025-07-17T10:13:55.8411395Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773 in _wrapped_call_impl 2025-07-17T10:13:55.8411924Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 803 in compile_wrapper 2025-07-17T10:13:55.8412424Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1784 in _call_impl 2025-07-17T10:13:55.8413003Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1773 in _wrapped_call_impl 2025-07-17T10:13:55.8413568Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/_dynamo/eval_frame.py", line 411 in __call__ 2025-07-17T10:13:55.8414018Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2625 in iter 2025-07-17T10:13:55.8414576Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees.py", line 2630 in test_skip_if_dynamic_shape_limit_reached2 2025-07-17T10:13:55.8415050Z File "/opt/conda/envs/py_3.12/lib/python3.12/contextlib.py", line 81 in inner 2025-07-17T10:13:55.8415502Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3223 in wrapper 2025-07-17T10:13:55.8415983Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 589 in _callTestMethod 2025-07-17T10:13:55.8416358Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 634 in run 2025-07-17T10:13:55.8416810Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3375 in _run_custom 2025-07-17T10:13:55.8417422Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 3403 in run 2025-07-17T10:13:55.8417873Z File "/opt/conda/envs/py_3.12/lib/python3.12/unittest/case.py", line 690 in __call__ 2025-07-17T10:13:55.8418281Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/unittest.py", line 333 in runtest 2025-07-17T10:13:55.8418751Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 169 in pytest_runtest_call 2025-07-17T10:13:55.8419219Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:13:55.8419665Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:13:55.8420101Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:13:55.8420537Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 262 in 2025-07-17T10:13:55.8420968Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 341 in from_call 2025-07-17T10:13:55.8421422Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 261 in call_runtest_hook 2025-07-17T10:13:55.8421895Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 222 in call_and_report 2025-07-17T10:13:55.8422355Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/runner.py", line 133 in runtestprotocol 2025-07-17T10:13:55.8422865Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pytest_rerunfailures.py", line 549 in pytest_runtest_protocol 2025-07-17T10:13:55.8423364Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:13:55.8423807Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:13:55.8424244Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:13:55.8424691Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 348 in pytest_runtestloop 2025-07-17T10:13:55.8425152Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:13:55.8425591Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:13:55.8426020Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:13:55.8426445Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 323 in _main 2025-07-17T10:13:55.8426870Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 269 in wrap_session 2025-07-17T10:13:55.8427390Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/main.py", line 316 in pytest_cmdline_main 2025-07-17T10:13:55.8427912Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_callers.py", line 121 in _multicall 2025-07-17T10:13:55.8428358Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_manager.py", line 120 in _hookexec 2025-07-17T10:13:55.8428838Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/pluggy/_hooks.py", line 512 in __call__ 2025-07-17T10:13:55.8429282Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/_pytest/config/__init__.py", line 166 in main 2025-07-17T10:13:55.8429797Z File "/opt/conda/envs/py_3.12/lib/python3.12/site-packages/torch/testing/_internal/common_utils.py", line 1298 in run_tests 2025-07-17T10:13:55.8430346Z File "/var/lib/jenkins/pytorch/test/inductor/test_cudagraph_trees_expandable_segments.py", line 40 in 2025-07-17T10:13:55.8430646Z 2025-07-17T10:13:55.8433844Z Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._bounded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards, torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, multidict._multidict, yarl._quoting_c, propcache._helpers_c, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket.mask, aiohttp._websocket.reader_c, frozenlist._frozenlist, charset_normalizer.md, requests.packages.charset_normalizer.md, requests.packages.chardet.md, thriftpy2.transport.cybase, thriftpy2.transport.buffered.cybuffered, thriftpy2.transport.framed.cyframed, thriftpy2.transport.memory.cymemory, thriftpy2.transport.sasl.cysasl, thriftpy2.protocol.cybin.cybin, scipy._lib._ccallback_c, yaml._yaml, numba.core.typeconv._typeconv, numba._helperlib, numba._dynfunc, numba._dispatcher, numba.core.runtime._nrt_python, numba.np.ufunc._internal, numba.experimental.jitclass._box, numba.mviewbuf, numba.types.itertools, markupsafe._speedups, psutil._psutil_linux, psutil._psutil_posix, hip_utils, __triton_launcher (total: 57) 2025-07-17T10:13:55.8437279Z Got exit code -6 (SIGIOT) 2025-07-17T10:13:55.8437596Z Test failed consistently, continuing with the rest of the tests due to continue-through-error being set 2025-07-17T10:13:55.8438561Z /opt/conda/envs/py_3.12/lib/python3.12/site-packages/hypothesis/entry_points.py:23: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-07-17T10:13:55.8439369Z import pkg_resources 2025-07-17T10:13:55.8439733Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to ./.pytorch-disabled-tests.json 2025-07-17T10:13:55.8440447Z Test results will be stored in test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-07c8f8cf6e9886dd.xml 2025-07-17T10:13:55.8440984Z ============================= test session starts ============================== 2025-07-17T10:13:55.8441324Z platform linux -- Python 3.12.11, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.12/bin/python 2025-07-17T10:13:55.8441622Z cachedir: .pytest_cache 2025-07-17T10:13:55.8441983Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-07-17T10:13:55.8442373Z rootdir: /var/lib/jenkins/pytorch 2025-07-17T10:13:55.8442560Z configfile: pytest.ini 2025-07-17T10:13:55.8443003Z plugins: subtests-0.13.1, xdist-3.3.1, rerunfailures-14.0, xdoctest-1.1.0, flakefinder-1.1.0, hypothesis-5.35.1, cpp-2.3.0, typeguard-4.3.0 2025-07-17T10:13:55.8443519Z collecting ... collected 129 items / 113 deselected / 16 selected 2025-07-17T10:13:55.8443782Z stepcurrent: skipping 113 already run items. 2025-07-17T10:13:55.8443988Z Running 16 items in this shard 2025-07-17T10:13:55.8444107Z 2025-07-17T10:13:55.8444535Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_symbolic <- test/inductor/test_cudagraph_trees.py PASSED [3.0333s] [ 6%] 2025-07-17T10:13:55.8445257Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_sparsity <- test/inductor/test_cudagraph_trees.py PASSED [0.4341s] [ 12%] 2025-07-17T10:13:55.8446020Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_static_inputs_address_mutation_log <- test/inductor/test_cudagraph_trees.py PASSED [2.3121s] [ 18%] 2025-07-17T10:13:55.8446805Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_storage_access_error <- test/inductor/test_cudagraph_trees.py PASSED [0.1774s] [ 25%] 2025-07-17T10:13:55.8447641Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_tensor_constant_mutation <- test/inductor/test_cudagraph_trees.py PASSED [1.0101s] [ 31%] 2025-07-17T10:13:55.8448425Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_tensor_dies_between_checkpoint <- test/inductor/test_cudagraph_trees.py PASSED [0.6475s] [ 37%] 2025-07-17T10:13:55.8449194Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_tensor_no_longer_in_pool <- test/inductor/test_cudagraph_trees.py PASSED [0.8054s] [ 43%] 2025-07-17T10:13:55.8449983Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_unaligned_static_input_no_cudagraphs <- test/inductor/test_cudagraph_trees.py PASSED [0.9388s] [ 50%] 2025-07-17T10:13:55.8450791Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_unaligned_static_input_non_trees <- test/inductor/test_cudagraph_trees.py PASSED [0.7992s] [ 56%] 2025-07-17T10:13:55.8451598Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_unaligned_static_input_trees <- test/inductor/test_cudagraph_trees.py PASSED [0.9390s] [ 62%] 2025-07-17T10:13:55.8452381Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_unaligned_static_parameter <- test/inductor/test_cudagraph_trees.py PASSED [0.4946s] [ 68%] 2025-07-17T10:13:55.8453131Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_unstable_ptr <- test/inductor/test_cudagraph_trees.py PASSED [0.6065s] [ 75%] 2025-07-17T10:13:55.8453861Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_warmup_stream_sync <- test/inductor/test_cudagraph_trees.py PASSED [0.5381s] [ 81%] 2025-07-17T10:13:55.8454612Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_warn_on_pending_backward <- test/inductor/test_cudagraph_trees.py PASSED [0.8545s] [ 87%] 2025-07-17T10:13:55.8455410Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_warn_once_if_dynamic_shape_limit_reached <- test/inductor/test_cudagraph_trees.py PASSED [5.0729s] [ 93%] 2025-07-17T10:13:55.8456366Z inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_workspace_allocation_error <- test/inductor/test_cudagraph_trees.py SKIPPED [0.1637s] (skipIfRocm: test doesn't currently work on the ROCm stack) [100%] 2025-07-17T10:13:55.8456919Z 2025-07-17T10:13:55.8457404Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_cudagraph_trees_expandable_segments/inductor.test_cudagraph_trees_expandable_segments-07c8f8cf6e9886dd.xml - 2025-07-17T10:13:55.8458046Z ================ 15 passed, 1 skipped, 113 deselected in 18.88s ================ 2025-07-17T10:13:55.8458580Z The following tests failed and then succeeded when run in a new process['test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_frozen_fn'] 2025-07-17T10:13:55.8459420Z The following tests failed consistently: ['test/inductor/test_cudagraph_trees_expandable_segments.py::CudaGraphTreeTests::test_skip_if_dynamic_shape_limit_reached2'] 2025-07-17T10:13:55.8459883Z 2025-07-17T10:13:55.8460274Z FINISHED PRINTING LOG FILE of inductor/test_cudagraph_trees_expandable_segments 1/1 (test/test-reports/inductor.test_cudagraph_trees_expandable_segments_1.1_e90ccb86446483a7_.log) 2025-07-17T10:13:55.8460711Z 2025-07-17T10:13:55.8460922Z GITHUB_RUN_ID, GITHUB_RUN_ATTEMPT, or ARTIFACTS_FILE_SUFFIX not set, not uploading 2025-07-17T10:13:55.8461280Z Running inductor/test_dependencies 1/1 ... [2025-07-17 10:13:55.800198] 2025-07-17T10:13:55.8461549Z Uploading artifacts took 0.00 seconds 2025-07-17T10:13:55.8461746Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:13:55.8461998Z inductor/test_cudagraph_trees_expandable_segments 1/1 failed! 2025-07-17T10:13:55.8462709Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_dependencies.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:13:55.800562] 2025-07-17T10:14:03.5831284Z 2025-07-17T10:14:03.5832429Z inductor/test_dependencies 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_dependencies_1.1_0923f107395b69e9_.log 2025-07-17T10:14:03.5834144Z Running 5 items in this shard: test/inductor/test_dependencies.py::TestDependencies::test_bucketize_dependencies_no_sorter, test/inductor/test_dependencies.py::TestDependencies::test_bucketize_dependencies_sorter, test/inductor/test_dependencies.py::TestDependencies::test_get_offset, test/inductor/test_dependencies.py::TestDependencies::test_normalize_with_stride_order_equal, test/inductor/test_dependencies.py::TestDependencies::test_normalize_with_stride_order_unequal 2025-07-17T10:14:03.5835539Z 2025-07-17T10:14:03.5835727Z Running inductor/test_efficient_conv_bn_eval 1/1 ... [2025-07-17 10:14:03.583045] 2025-07-17T10:14:03.5836037Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:14:03.5836732Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_efficient_conv_bn_eval.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:14:03.583383] 2025-07-17T10:14:11.2651101Z 2025-07-17T10:14:11.2651889Z inductor/test_efficient_conv_bn_eval 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_efficient_conv_bn_eval_1.1_6045dc307fa87b6e_.log 2025-07-17T10:14:11.2652872Z Running 2 items in this shard: test/inductor/test_efficient_conv_bn_eval.py::EfficientConvBNEvalCpuTests::test_basic_cpu, test/inductor/test_efficient_conv_bn_eval.py::EfficientConvBNEvalGpuTests::test_basic_cuda 2025-07-17T10:14:11.2653381Z 2025-07-17T10:14:11.2653553Z Running inductor/test_extension_backend 1/1 ... [2025-07-17 10:14:11.265101] 2025-07-17T10:14:11.2653852Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:14:11.2655877Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_extension_backend.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:14:11.265442] 2025-07-17T10:14:15.9513581Z 2025-07-17T10:14:15.9514472Z inductor/test_decompose_mem_bound_mm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_decompose_mem_bound_mm_1.1_7dfc28cce3d1d5ab_.log 2025-07-17T10:14:15.9526426Z Running 36 items in this shard: test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_check_device, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_b_10240_m_2_k_2_n_2_should_decompose_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_b_10240_m_2_k_32_n_32_should_decompose_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_b_2000_m_2_k_2_n_2_should_decompose_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_cpu_b_1_m_2_k_2_n_2_should_decompose_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_bmm_cpu_b_2_m_2_k_2_n_2_should_decompose_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_20480_k_32_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_20480_k_32_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_20480_k_5_n_2_should_decompose_True_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_20480_k_5_n_2_should_decompose_True_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_2048_k_2_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_m_2048_k_2_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_32_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_32_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_2048_k_2_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_linear_mixed_precision_m_2048_k_2_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_cpu_m_1_k_64_n_16_should_decompose_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_cpu_m_1_k_64_n_32_should_decompose_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_cpu_m_2_k_64_n_16_should_decompose_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_20480_k_32_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_20480_k_32_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_20480_k_5_n_2_should_decompose_True_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_20480_k_5_n_2_should_decompose_True_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_2048_k_2_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_m_2048_k_2_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_32_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_32_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_20480_k_5_n_2_should_decompose_True_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_2048_k_2_n_2_should_decompose_False_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_decompose_mm_mixed_precision_m_2048_k_2_n_2_should_decompose_False_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_dynamic_shape_m_20480_k_5_n_2_should_decompose_True_has_bias_False, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_dynamic_shape_m_20480_k_5_n_2_should_decompose_True_has_bias_True, test/inductor/test_decompose_mem_bound_mm.py::TestDecomposeMemMM::test_realize_input 2025-07-17T10:14:15.9537572Z 2025-07-17T10:14:15.9537730Z Running inductor/test_external_callables 1/1 ... [2025-07-17 10:14:15.951340] 2025-07-17T10:14:15.9538032Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:14:15.9538728Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_external_callables.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:14:15.951716] 2025-07-17T10:14:35.5097255Z 2025-07-17T10:14:35.5098121Z inductor/test_external_callables 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_external_callables_1.1_04b26e35e9c20064_.log 2025-07-17T10:14:35.5099349Z Running 3 items in this shard: test/inductor/test_external_callables.py::TestInductorExternalCallable::test_matmul_cpu, test/inductor/test_external_callables.py::TestInductorExternalCallable::test_matmul_cuda, test/inductor/test_external_callables.py::TestInductorExternalCallable::test_matmul_dup 2025-07-17T10:14:35.5100184Z 2025-07-17T10:14:35.5100349Z Running inductor/test_flex_decoding 1/3 ... [2025-07-17 10:14:35.509637] 2025-07-17T10:14:35.5100633Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:14:35.5101312Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_flex_decoding.py', '-m', 'not serial', '--shard-id=1', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:14:35.509957] 2025-07-17T10:14:37.1882012Z 2025-07-17T10:14:37.1882914Z inductor/test_extension_backend 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_extension_backend_1.1_3a49f159eef5542b_.log 2025-07-17T10:14:37.1883697Z Running 1 items in this shard: test/inductor/test_extension_backend.py::ExtensionBackendTests::test_open_device_registration 2025-07-17T10:14:37.1884042Z 2025-07-17T10:14:37.1890010Z Running inductor/test_mps_basic 1/1 ... [2025-07-17 10:14:37.188069] 2025-07-17T10:14:37.1890317Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:14:37.1891024Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_mps_basic.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:14:37.188379] 2025-07-17T10:14:44.6730114Z 2025-07-17T10:14:44.6730898Z inductor/test_mps_basic 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_mps_basic_1.1_52ef3df5425d4dcc_.log 2025-07-17T10:14:44.6731324Z 2025-07-17T10:14:44.6731509Z Running inductor/test_ordered_set 1/1 ... [2025-07-17 10:14:44.672890] 2025-07-17T10:14:44.6731792Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:14:44.6733577Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_ordered_set.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:14:44.673207] 2025-07-17T10:14:49.7989854Z 2025-07-17T10:14:49.7991337Z inductor/test_ordered_set 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_ordered_set_1.1_0bd1b0de88cc870c_.log 2025-07-17T10:14:49.8060366Z Running 401 items in this shard: test/inductor/test_ordered_set.py::TestJointOps::test_and, test/inductor/test_ordered_set.py::TestJointOps::test_badcmp, test/inductor/test_ordered_set.py::TestJointOps::test_container_iterator, test/inductor/test_ordered_set.py::TestJointOps::test_contains, test/inductor/test_ordered_set.py::TestJointOps::test_cyclical_repr, test/inductor/test_ordered_set.py::TestJointOps::test_deepcopy, test/inductor/test_ordered_set.py::TestJointOps::test_difference, test/inductor/test_ordered_set.py::TestJointOps::test_do_not_rehash_dict_keys, test/inductor/test_ordered_set.py::TestJointOps::test_equality, test/inductor/test_ordered_set.py::TestJointOps::test_free_after_iterating, test/inductor/test_ordered_set.py::TestJointOps::test_gc, test/inductor/test_ordered_set.py::TestJointOps::test_intersection, test/inductor/test_ordered_set.py::TestJointOps::test_isdisjoint, test/inductor/test_ordered_set.py::TestJointOps::test_iterator_pickling, test/inductor/test_ordered_set.py::TestJointOps::test_len, test/inductor/test_ordered_set.py::TestJointOps::test_new_or_init, test/inductor/test_ordered_set.py::TestJointOps::test_or, test/inductor/test_ordered_set.py::TestJointOps::test_pickling, test/inductor/test_ordered_set.py::TestJointOps::test_setOfFrozensets, test/inductor/test_ordered_set.py::TestJointOps::test_sub, test/inductor/test_ordered_set.py::TestJointOps::test_sub_and_super, test/inductor/test_ordered_set.py::TestJointOps::test_subclass_with_custom_hash, test/inductor/test_ordered_set.py::TestJointOps::test_symmetric_difference, test/inductor/test_ordered_set.py::TestJointOps::test_union, test/inductor/test_ordered_set.py::TestJointOps::test_uniquification, test/inductor/test_ordered_set.py::TestJointOps::test_xor, test/inductor/test_ordered_set.py::TestSet::test_add, test/inductor/test_ordered_set.py::TestSet::test_and, test/inductor/test_ordered_set.py::TestSet::test_badcmp, test/inductor/test_ordered_set.py::TestSet::test_clear, test/inductor/test_ordered_set.py::TestSet::test_constructor_identity, test/inductor/test_ordered_set.py::TestSet::test_container_iterator, test/inductor/test_ordered_set.py::TestSet::test_contains, test/inductor/test_ordered_set.py::TestSet::test_copy, test/inductor/test_ordered_set.py::TestSet::test_cyclical_repr, test/inductor/test_ordered_set.py::TestSet::test_deepcopy, test/inductor/test_ordered_set.py::TestSet::test_difference, test/inductor/test_ordered_set.py::TestSet::test_difference_update, test/inductor/test_ordered_set.py::TestSet::test_discard, test/inductor/test_ordered_set.py::TestSet::test_do_not_rehash_dict_keys, test/inductor/test_ordered_set.py::TestSet::test_equality, test/inductor/test_ordered_set.py::TestSet::test_free_after_iterating, test/inductor/test_ordered_set.py::TestSet::test_gc, test/inductor/test_ordered_set.py::TestSet::test_hash, test/inductor/test_ordered_set.py::TestSet::test_iand, test/inductor/test_ordered_set.py::TestSet::test_init, test/inductor/test_ordered_set.py::TestSet::test_inplace_on_self, test/inductor/test_ordered_set.py::TestSet::test_intersection, test/inductor/test_ordered_set.py::TestSet::test_intersection_update, test/inductor/test_ordered_set.py::TestSet::test_ior, test/inductor/test_ordered_set.py::TestSet::test_isdisjoint, test/inductor/test_ordered_set.py::TestSet::test_isub, test/inductor/test_ordered_set.py::TestSet::test_iterator_pickling, test/inductor/test_ordered_set.py::TestSet::test_ixor, test/inductor/test_ordered_set.py::TestSet::test_len, test/inductor/test_ordered_set.py::TestSet::test_new_or_init, test/inductor/test_ordered_set.py::TestSet::test_or, test/inductor/test_ordered_set.py::TestSet::test_pickling, test/inductor/test_ordered_set.py::TestSet::test_pop, test/inductor/test_ordered_set.py::TestSet::test_remove, test/inductor/test_ordered_set.py::TestSet::test_remove_keyerror_set, test/inductor/test_ordered_set.py::TestSet::test_remove_keyerror_unpacking, test/inductor/test_ordered_set.py::TestSet::test_rich_compare, test/inductor/test_ordered_set.py::TestSet::test_setOfFrozensets, test/inductor/test_ordered_set.py::TestSet::test_set_literal, test/inductor/test_ordered_set.py::TestSet::test_set_literal_evaluation_order, test/inductor/test_ordered_set.py::TestSet::test_set_literal_insertion_order, test/inductor/test_ordered_set.py::TestSet::test_sub, test/inductor/test_ordered_set.py::TestSet::test_sub_and_super, test/inductor/test_ordered_set.py::TestSet::test_subclass_with_custom_hash, test/inductor/test_ordered_set.py::TestSet::test_symmetric_difference, test/inductor/test_ordered_set.py::TestSet::test_symmetric_difference_update, test/inductor/test_ordered_set.py::TestSet::test_union, test/inductor/test_ordered_set.py::TestSet::test_uniquification, test/inductor/test_ordered_set.py::TestSet::test_update, test/inductor/test_ordered_set.py::TestSet::test_weakref, test/inductor/test_ordered_set.py::TestSet::test_xor, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_copy, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_empty_difference, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_empty_difference_rev, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_empty_intersection, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_empty_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_empty_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_empty_union, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_equivalent_equality, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_intersection_empty, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_isdisjoint_empty, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_issue_37219, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_iteration, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_length, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_pickling, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_repr, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_self_difference, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_self_equality, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_self_intersection, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_self_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_self_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_self_union, test/inductor/test_ordered_set.py::TestBasicOpsEmpty::test_union_empty, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_copy, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_empty_difference, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_empty_difference_rev, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_empty_intersection, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_empty_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_empty_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_empty_union, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_equivalent_equality, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_in, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_intersection_empty, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_isdisjoint_empty, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_issue_37219, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_iteration, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_length, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_not_in, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_pickling, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_repr, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_self_difference, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_self_equality, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_self_intersection, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_self_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_self_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_self_union, test/inductor/test_ordered_set.py::TestBasicOpsSingleton::test_union_empty, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_copy, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_empty_difference, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_empty_difference_rev, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_empty_intersection, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_empty_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_empty_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_empty_union, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_equivalent_equality, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_in, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_intersection_empty, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_isdisjoint_empty, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_issue_37219, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_iteration, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_length, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_not_in, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_pickling, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_repr, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_self_difference, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_self_equality, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_self_intersection, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_self_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_self_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_self_union, test/inductor/test_ordered_set.py::TestBasicOpsTuple::test_union_empty, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_copy, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_empty_difference, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_empty_difference_rev, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_empty_intersection, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_empty_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_empty_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_empty_union, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_equivalent_equality, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_intersection_empty, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_isdisjoint_empty, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_issue_37219, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_iteration, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_length, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_pickling, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_repr, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_self_difference, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_self_equality, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_self_intersection, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_self_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_self_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_self_union, test/inductor/test_ordered_set.py::TestBasicOpsTriple::test_union_empty, test/inductor/test_ordered_set.py::TestBasicOpsString::test_copy, test/inductor/test_ordered_set.py::TestBasicOpsString::test_empty_difference, test/inductor/test_ordered_set.py::TestBasicOpsString::test_empty_difference_rev, test/inductor/test_ordered_set.py::TestBasicOpsString::test_empty_intersection, test/inductor/test_ordered_set.py::TestBasicOpsString::test_empty_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsString::test_empty_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsString::test_empty_union, test/inductor/test_ordered_set.py::TestBasicOpsString::test_equivalent_equality, test/inductor/test_ordered_set.py::TestBasicOpsString::test_intersection_empty, test/inductor/test_ordered_set.py::TestBasicOpsString::test_isdisjoint_empty, test/inductor/test_ordered_set.py::TestBasicOpsString::test_issue_37219, test/inductor/test_ordered_set.py::TestBasicOpsString::test_iteration, test/inductor/test_ordered_set.py::TestBasicOpsString::test_length, test/inductor/test_ordered_set.py::TestBasicOpsString::test_pickling, test/inductor/test_ordered_set.py::TestBasicOpsString::test_repr, test/inductor/test_ordered_set.py::TestBasicOpsString::test_self_difference, test/inductor/test_ordered_set.py::TestBasicOpsString::test_self_equality, test/inductor/test_ordered_set.py::TestBasicOpsString::test_self_intersection, test/inductor/test_ordered_set.py::TestBasicOpsString::test_self_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsString::test_self_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsString::test_self_union, test/inductor/test_ordered_set.py::TestBasicOpsString::test_union_empty, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_copy, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_empty_difference, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_empty_difference_rev, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_empty_intersection, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_empty_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_empty_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_empty_union, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_equivalent_equality, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_intersection_empty, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_isdisjoint_empty, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_issue_37219, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_iteration, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_length, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_pickling, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_repr, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_self_difference, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_self_equality, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_self_intersection, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_self_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_self_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_self_union, test/inductor/test_ordered_set.py::TestBasicOpsBytes::test_union_empty, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_copy, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_empty_difference, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_empty_difference_rev, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_empty_intersection, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_empty_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_empty_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_empty_union, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_equivalent_equality, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_intersection_empty, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_isdisjoint_empty, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_issue_37219, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_iteration, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_length, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_pickling, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_repr, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_self_difference, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_self_equality, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_self_intersection, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_self_isdisjoint, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_self_symmetric_difference, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_self_union, test/inductor/test_ordered_set.py::TestBasicOpsMixedStringBytes::test_union_empty, test/inductor/test_ordered_set.py::TestExceptionPropagation::test_changingSizeWhileIterating, test/inductor/test_ordered_set.py::TestExceptionPropagation::test_instanceWithException, test/inductor/test_ordered_set.py::TestExceptionPropagation::test_instancesWithoutException, test/inductor/test_ordered_set.py::TestSetOfSets::test_constructor, test/inductor/test_ordered_set.py::TestBinaryOps::test_eq, test/inductor/test_ordered_set.py::TestBinaryOps::test_intersection_non_overlap, test/inductor/test_ordered_set.py::TestBinaryOps::test_intersection_overlap, test/inductor/test_ordered_set.py::TestBinaryOps::test_intersection_subset, test/inductor/test_ordered_set.py::TestBinaryOps::test_intersection_superset, test/inductor/test_ordered_set.py::TestBinaryOps::test_isdisjoint_non_overlap, test/inductor/test_ordered_set.py::TestBinaryOps::test_isdisjoint_overlap, test/inductor/test_ordered_set.py::TestBinaryOps::test_isdisjoint_subset, test/inductor/test_ordered_set.py::TestBinaryOps::test_isdisjoint_superset, test/inductor/test_ordered_set.py::TestBinaryOps::test_sym_difference_non_overlap, test/inductor/test_ordered_set.py::TestBinaryOps::test_sym_difference_overlap, test/inductor/test_ordered_set.py::TestBinaryOps::test_sym_difference_subset, test/inductor/test_ordered_set.py::TestBinaryOps::test_sym_difference_superset, test/inductor/test_ordered_set.py::TestBinaryOps::test_union_non_overlap, test/inductor/test_ordered_set.py::TestBinaryOps::test_union_overlap, test/inductor/test_ordered_set.py::TestBinaryOps::test_union_subset, test/inductor/test_ordered_set.py::TestBinaryOps::test_union_superset, test/inductor/test_ordered_set.py::TestUpdateOps::test_difference_method_call, test/inductor/test_ordered_set.py::TestUpdateOps::test_difference_non_overlap, test/inductor/test_ordered_set.py::TestUpdateOps::test_difference_overlap, test/inductor/test_ordered_set.py::TestUpdateOps::test_difference_subset, test/inductor/test_ordered_set.py::TestUpdateOps::test_difference_superset, test/inductor/test_ordered_set.py::TestUpdateOps::test_intersection_method_call, test/inductor/test_ordered_set.py::TestUpdateOps::test_intersection_non_overlap, test/inductor/test_ordered_set.py::TestUpdateOps::test_intersection_overlap, test/inductor/test_ordered_set.py::TestUpdateOps::test_intersection_subset, test/inductor/test_ordered_set.py::TestUpdateOps::test_intersection_superset, test/inductor/test_ordered_set.py::TestUpdateOps::test_sym_difference_method_call, test/inductor/test_ordered_set.py::TestUpdateOps::test_sym_difference_non_overlap, test/inductor/test_ordered_set.py::TestUpdateOps::test_sym_difference_overlap, test/inductor/test_ordered_set.py::TestUpdateOps::test_sym_difference_subset, test/inductor/test_ordered_set.py::TestUpdateOps::test_sym_difference_superset, test/inductor/test_ordered_set.py::TestUpdateOps::test_union_method_call, test/inductor/test_ordered_set.py::TestUpdateOps::test_union_non_overlap, test/inductor/test_ordered_set.py::TestUpdateOps::test_union_overlap, test/inductor/test_ordered_set.py::TestUpdateOps::test_union_subset, test/inductor/test_ordered_set.py::TestUpdateOps::test_union_superset, test/inductor/test_ordered_set.py::TestMutate::test_add_absent, test/inductor/test_ordered_set.py::TestMutate::test_add_present, test/inductor/test_ordered_set.py::TestMutate::test_add_until_full, test/inductor/test_ordered_set.py::TestMutate::test_clear, test/inductor/test_ordered_set.py::TestMutate::test_discard_absent, test/inductor/test_ordered_set.py::TestMutate::test_discard_present, test/inductor/test_ordered_set.py::TestMutate::test_pop, test/inductor/test_ordered_set.py::TestMutate::test_remove_absent, test/inductor/test_ordered_set.py::TestMutate::test_remove_present, test/inductor/test_ordered_set.py::TestMutate::test_remove_until_empty, test/inductor/test_ordered_set.py::TestMutate::test_update_empty_tuple, test/inductor/test_ordered_set.py::TestMutate::test_update_unit_tuple_non_overlap, test/inductor/test_ordered_set.py::TestMutate::test_update_unit_tuple_overlap, test/inductor/test_ordered_set.py::TestSubsets::test_issubset, test/inductor/test_ordered_set.py::TestSubsetEqualEmpty::test_issubset, test/inductor/test_ordered_set.py::TestSubsetEqualNonEmpty::test_issubset, test/inductor/test_ordered_set.py::TestSubsetEmptyNonEmpty::test_issubset, test/inductor/test_ordered_set.py::TestSubsetPartial::test_issubset, test/inductor/test_ordered_set.py::TestSubsetNonOverlap::test_issubset, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_difference, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_difference_update, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_difference_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_eq_ne, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_ge_gt_le_lt, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_intersection, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_intersection_update, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_intersection_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_sym_difference, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_sym_difference_update, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_sym_difference_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_union, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_update, test/inductor/test_ordered_set.py::TestOnlySetsNumeric::test_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_difference, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_difference_update, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_difference_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_eq_ne, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_ge_gt_le_lt, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_intersection, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_intersection_update, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_intersection_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_sym_difference, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_sym_difference_update, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_sym_difference_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_union, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_update, test/inductor/test_ordered_set.py::TestOnlySetsDict::test_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_difference, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_difference_update, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_difference_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_eq_ne, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_ge_gt_le_lt, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_intersection, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_intersection_update, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_intersection_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_sym_difference, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_sym_difference_update, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_sym_difference_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_union, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_update, test/inductor/test_ordered_set.py::TestOnlySetsOperator::test_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_difference, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_difference_update, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_difference_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_eq_ne, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_ge_gt_le_lt, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_intersection, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_intersection_update, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_intersection_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_sym_difference, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_sym_difference_update, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_sym_difference_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_union, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_update, test/inductor/test_ordered_set.py::TestOnlySetsTuple::test_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsString::test_difference, test/inductor/test_ordered_set.py::TestOnlySetsString::test_difference_update, test/inductor/test_ordered_set.py::TestOnlySetsString::test_difference_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsString::test_eq_ne, test/inductor/test_ordered_set.py::TestOnlySetsString::test_ge_gt_le_lt, test/inductor/test_ordered_set.py::TestOnlySetsString::test_intersection, test/inductor/test_ordered_set.py::TestOnlySetsString::test_intersection_update, test/inductor/test_ordered_set.py::TestOnlySetsString::test_intersection_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsString::test_sym_difference, test/inductor/test_ordered_set.py::TestOnlySetsString::test_sym_difference_update, test/inductor/test_ordered_set.py::TestOnlySetsString::test_sym_difference_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsString::test_union, test/inductor/test_ordered_set.py::TestOnlySetsString::test_update, test/inductor/test_ordered_set.py::TestOnlySetsString::test_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_difference, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_difference_update, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_difference_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_eq_ne, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_ge_gt_le_lt, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_intersection, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_intersection_update, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_intersection_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_sym_difference, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_sym_difference_update, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_sym_difference_update_operator, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_union, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_update, test/inductor/test_ordered_set.py::TestOnlySetsGenerator::test_update_operator, test/inductor/test_ordered_set.py::TestCopyingEmpty::test_copy, test/inductor/test_ordered_set.py::TestCopyingEmpty::test_deep_copy, test/inductor/test_ordered_set.py::TestCopyingSingleton::test_copy, test/inductor/test_ordered_set.py::TestCopyingSingleton::test_deep_copy, test/inductor/test_ordered_set.py::TestCopyingTriple::test_copy, test/inductor/test_ordered_set.py::TestCopyingTriple::test_deep_copy, test/inductor/test_ordered_set.py::TestCopyingTuple::test_copy, test/inductor/test_ordered_set.py::TestCopyingTuple::test_deep_copy, test/inductor/test_ordered_set.py::TestCopyingNested::test_copy, test/inductor/test_ordered_set.py::TestCopyingNested::test_deep_copy, test/inductor/test_ordered_set.py::TestIdentities::test_binopsVsSubsets, test/inductor/test_ordered_set.py::TestIdentities::test_commutativity, test/inductor/test_ordered_set.py::TestIdentities::test_exclusion, test/inductor/test_ordered_set.py::TestIdentities::test_summations, test/inductor/test_ordered_set.py::TestVariousIteratorArgs::test_constructor, test/inductor/test_ordered_set.py::TestVariousIteratorArgs::test_inline_methods, test/inductor/test_ordered_set.py::TestVariousIteratorArgs::test_inplace_methods, test/inductor/test_ordered_set.py::TestWeirdBugs::test_8420_set_merge, test/inductor/test_ordered_set.py::TestWeirdBugs::test_iter_and_mutate, test/inductor/test_ordered_set.py::TestWeirdBugs::test_merge_and_mutate, test/inductor/test_ordered_set.py::TestGraphs::test_cube, test/inductor/test_ordered_set.py::TestGraphs::test_cuboctahedron 2025-07-17T10:14:49.8133143Z 2025-07-17T10:14:49.8133277Z Running inductor/test_pad_mm 1/1 ... [2025-07-17 10:14:49.799330] 2025-07-17T10:14:49.8133549Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:14:49.8134203Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_pad_mm.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:14:49.799677] 2025-07-17T10:15:59.6190411Z 2025-07-17T10:15:59.6193722Z inductor/test_pad_mm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_pad_mm_1.1_7298b16df8e46069_.log 2025-07-17T10:15:59.6196653Z Running 18 items in this shard: test/inductor/test_pad_mm.py::PadMMTest::test_cat_pad_mm_dyn_m, test/inductor/test_pad_mm.py::PadMMTest::test_exclude_cat_padding, test/inductor/test_pad_mm.py::PadMMTest::test_exclude_padding, test/inductor/test_pad_mm.py::PadMMTest::test_original_aten_preserved_pad_mm, test/inductor/test_pad_mm.py::PadMMTest::test_pad_addmm_2d_bias, test/inductor/test_pad_mm.py::PadMMTest::test_pad_addmm_dyn_m, test/inductor/test_pad_mm.py::PadMMTest::test_pad_addmm_dyn_mn, test/inductor/test_pad_mm.py::PadMMTest::test_pad_batch, test/inductor/test_pad_mm.py::PadMMTest::test_pad_bmm_dyn_b, test/inductor/test_pad_mm.py::PadMMTest::test_pad_bmm_dyn_bm, test/inductor/test_pad_mm.py::PadMMTest::test_pad_bmm_dyn_k, test/inductor/test_pad_mm.py::PadMMTest::test_pad_mm_bf16, test/inductor/test_pad_mm.py::PadMMTest::test_pad_mm_dyn_k, test/inductor/test_pad_mm.py::PadMMTest::test_pad_mm_dyn_m, test/inductor/test_pad_mm.py::PadMMTest::test_pad_mm_dyn_mnk, test/inductor/test_pad_mm.py::PadMMTest::test_pad_mm_dyn_n, test/inductor/test_pad_mm.py::PadMMTest::test_pad_single_cat, test/inductor/test_pad_mm.py::PadMMTest::test_zero_dim 2025-07-17T10:15:59.6199222Z 2025-07-17T10:15:59.6199365Z Running inductor/test_padding 1/1 ... [2025-07-17 10:15:59.619021] 2025-07-17T10:15:59.6199641Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:15:59.6200308Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_padding.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:15:59.619374] 2025-07-17T10:16:34.1097336Z 2025-07-17T10:16:34.1098010Z inductor/test_padding 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_padding_1.1_7eb0b7f944180f3f_.log 2025-07-17T10:16:34.1105001Z Running 30 items in this shard: test/inductor/test_padding.py::PerfTestBetweenGoodAndBadShape::test_BertForMaskedLM, test/inductor/test_padding.py::PerfTestBetweenGoodAndBadShape::test_LinearAndSoftmax_both_shapes, test/inductor/test_padding.py::PerfTestBetweenGoodAndBadShape::test_nobias_LinearAndSoftmax_both_shapes, test/inductor/test_padding.py::PerfTestWithAndWithoutPadding::test_longformer, test/inductor/test_padding.py::PerfTestWithAndWithoutPadding::test_longformer_small_bs, test/inductor/test_padding.py::PerfTestWithAndWithoutPadding::test_nvidia_deeprecommender, test/inductor/test_padding.py::PaddingTest::test_LinearAndSoftmax_codegen, test/inductor/test_padding.py::PaddingTest::test_attention, test/inductor/test_padding.py::PaddingTest::test_cat, test/inductor/test_padding.py::PaddingTest::test_conv, test/inductor/test_padding.py::PaddingTest::test_matmul, test/inductor/test_padding.py::PaddingTest::test_mm_padding_perf, test/inductor/test_padding.py::PaddingTest::test_noop_concat_output_padding_shape0_alignment_bytes_32_pad_output_False, test/inductor/test_padding.py::PaddingTest::test_noop_concat_output_padding_shape1_alignment_bytes_32_pad_output_True, test/inductor/test_padding.py::PaddingTest::test_noop_concat_output_padding_shape2_alignment_bytes_64_pad_output_False, test/inductor/test_padding.py::PaddingTest::test_noop_concat_output_padding_shape3_alignment_bytes_64_pad_output_True, test/inductor/test_padding.py::PaddingTest::test_pad_3d_tensor, test/inductor/test_padding.py::PaddingTest::test_pad_channels_last, test/inductor/test_padding.py::PaddingTest::test_pad_outputs_alignment_bytes_128_shape0_float16, test/inductor/test_padding.py::PaddingTest::test_pad_outputs_alignment_bytes_128_shape0_float32, test/inductor/test_padding.py::PaddingTest::test_pad_outputs_alignment_bytes_128_shape1_float16, test/inductor/test_padding.py::PaddingTest::test_pad_outputs_alignment_bytes_128_shape1_float32, test/inductor/test_padding.py::PaddingTest::test_pad_outputs_alignment_bytes_32_shape0_float16, test/inductor/test_padding.py::PaddingTest::test_pad_outputs_alignment_bytes_32_shape0_float32, test/inductor/test_padding.py::PaddingTest::test_pad_outputs_alignment_bytes_32_shape1_float16, test/inductor/test_padding.py::PaddingTest::test_pad_outputs_alignment_bytes_32_shape1_float32, test/inductor/test_padding.py::PaddingTest::test_pad_strides, test/inductor/test_padding.py::PaddingTest::test_pad_strides_skip, test/inductor/test_padding.py::PaddingTest::test_padmm, test/inductor/test_padding.py::PaddingTest::test_view 2025-07-17T10:16:34.1111325Z 2025-07-17T10:16:34.1111475Z Running inductor/test_pattern_matcher 1/1 ... [2025-07-17 10:16:34.109659] 2025-07-17T10:16:34.1111763Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:16:34.1112443Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_pattern_matcher.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:16:34.110012] 2025-07-17T10:17:24.9477117Z 2025-07-17T10:17:24.9480466Z inductor/test_pattern_matcher 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_pattern_matcher_1.1_4ba8a545c94ef1e4_.log 2025-07-17T10:17:24.9490660Z Running 45 items in this shard: test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_addmm, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_addmm_broadcasting_bias, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_addmm_symbolic_scalar, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_bmm_to_mm, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_cat_addmm, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_cat_mm, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_cat_slice_cat_cuda, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_cat_splitwithsizes, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_duplicate_search, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_fused_int_mm_mul, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_fused_int_mm_mul_epilogue, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_fused_int_mm_mul_gating, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_match_equivalent_function_invocations1, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_match_equivalent_function_invocations2, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_match_equivalent_function_invocations3, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_match_with_mutation, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_mixed_mm, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_mixed_mm_bad_cases, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_mixed_mm_cpu, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_mixed_mm_epi_works, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_mixed_mm_exhaustive_dtypes, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_mixed_mm_gating, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_mm_plus_mm, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_multioutput_register_replacement, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_mutation_op_matching, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_original_aten_preserved_split_addmm, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_pointless_convert, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_pointless_cumsum, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_pointless_permute_pair, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_pointless_permute_pair_3d, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_pointless_view_pair, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_pointless_view_pair_dynamic_shapes, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_remove_pointless_clones, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_replace_mul_zero, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_scaled_softmax, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_serialized_patterns_up_to_date, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_splitwithsizes_cat, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_stable_topological_sort, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_successful_partial_reuse_case0, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_successful_partial_reuse_case1, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_successful_partial_reuse_case2, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_symint_pattern_matching, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_unfuse_bias_addmm, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_unsuccessful_partial_reuse_case0, test/inductor/test_pattern_matcher.py::TestPatternMatcher::test_unsuccessful_partial_reuse_case1 2025-07-17T10:17:24.9500146Z 2025-07-17T10:17:24.9500283Z Running inductor/test_perf 1/1 ... [2025-07-17 10:17:24.947621] 2025-07-17T10:17:24.9500563Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:17:24.9501206Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_perf.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:17:24.947944] 2025-07-17T10:18:19.9862902Z 2025-07-17T10:18:19.9863610Z inductor/test_perf 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_perf_1.1_ed47bbb1b2291ff1_.log 2025-07-17T10:18:19.9875835Z Running 66 items in this shard: test/inductor/test_perf.py::NumBytesMetricTests::test_cat, test/inductor/test_perf.py::NumBytesMetricTests::test_cat_pointwise, test/inductor/test_perf.py::NumBytesMetricTests::test_cat_pointwise_config_option, test/inductor/test_perf.py::NumBytesMetricTests::test_cat_pointwise_many_complex_inputs, test/inductor/test_perf.py::NumBytesMetricTests::test_cat_pointwise_many_simple_inputs, test/inductor/test_perf.py::NumBytesMetricTests::test_extern, test/inductor/test_perf.py::NumBytesMetricTests::test_index, test/inductor/test_perf.py::NumBytesMetricTests::test_pointwise, test/inductor/test_perf.py::NumBytesMetricTests::test_reduction, test/inductor/test_perf.py::FusionTests::test_create_block_mask, test/inductor/test_perf.py::FusionTests::test_double_softmax, test/inductor/test_perf.py::FusionTests::test_factory_reduction, test/inductor/test_perf.py::FusionTests::test_horizontal_reduction_outer_pointwise, test/inductor/test_perf.py::FusionTests::test_horizontal_reduction_pointwise, test/inductor/test_perf.py::FusionTests::test_horizontal_reduction_pointwise2, test/inductor/test_perf.py::FusionTests::test_horizontal_reduction_reduction, test/inductor/test_perf.py::FusionTests::test_horizontal_sum_pw_broadcast, test/inductor/test_perf.py::FusionTests::test_index_pointwise, test/inductor/test_perf.py::FusionTests::test_index_reduction, test/inductor/test_perf.py::FusionTests::test_layer_norm, test/inductor/test_perf.py::FusionTests::test_mutation_fusion, test/inductor/test_perf.py::FusionTests::test_neighbor, test/inductor/test_perf.py::FusionTests::test_norm_chain, test/inductor/test_perf.py::FusionTests::test_pointwise_multi_level_reduction, test/inductor/test_perf.py::FusionTests::test_reduction_pointwise_multi_level_reduction, test/inductor/test_perf.py::FusionTests::test_softmax_backward, test/inductor/test_perf.py::FusionTests::test_softmax_inner, test/inductor/test_perf.py::FusionTests::test_vertical_sum_pw, test/inductor/test_perf.py::SchedulerFusionTests::test_fusion_choice1, test/inductor/test_perf.py::SchedulerFusionTests::test_fusion_choice2, test/inductor/test_perf.py::SchedulerFusionTests::test_fusion_choice3, test/inductor/test_perf.py::SchedulerFusionTests::test_fusion_choice4_cpu, test/inductor/test_perf.py::TilingTests::test_tiling_simple, test/inductor/test_perf.py::TilingTests::test_tiling_three, test/inductor/test_perf.py::MinCutPartitioningTests::test_partitioning_cat, test/inductor/test_perf.py::MinCutPartitioningTests::test_partitioning_dtype, test/inductor/test_perf.py::MinCutPartitioningTests::test_partitioning_full_remat, test/inductor/test_perf.py::MinCutPartitioningTests::test_partitioning_keops, test/inductor/test_perf.py::MinCutPartitioningTests::test_partitioning_long_chain_add, test/inductor/test_perf.py::MinCutPartitioningTests::test_partitioning_partial_remat, test/inductor/test_perf.py::MinCutPartitioningTests::test_partitioning_relu, test/inductor/test_perf.py::MinCutPartitioningTests::test_partitioning_unremat_bw, test/inductor/test_perf.py::MinCutPartitioningTests::test_partitioning_unremat_bw2, test/inductor/test_perf.py::MinCutPartitioningTests::test_partitioning_with_view, test/inductor/test_perf.py::NoopTests::test_noop_cat, test/inductor/test_perf.py::NoopTests::test_noop_clones, test/inductor/test_perf.py::NoopTests::test_noop_device_conversion, test/inductor/test_perf.py::NoopTests::test_noop_dtype_conversion, test/inductor/test_perf.py::NoopTests::test_noop_int_ops, test/inductor/test_perf.py::NoopTests::test_noop_slice_scatter, test/inductor/test_perf.py::InplacingTests::test_inplace_custom_op, test/inductor/test_perf.py::InplacingTests::test_inplace_custom_op_intermediate, test/inductor/test_perf.py::InplacingTests::test_inplace_custom_op_training, test/inductor/test_perf.py::InplacingTests::test_inplace_custom_op_training_two_mutated_inputs, test/inductor/test_perf.py::InplacingTests::test_inplace_custom_op_two_mutated_inputs, test/inductor/test_perf.py::InplacingTests::test_inplace_randperm_scatter, test/inductor/test_perf.py::InplacingTests::test_inplace_scatter, test/inductor/test_perf.py::InplacingTests::test_inplace_scatter_noop_view, test/inductor/test_perf.py::InplacingTests::test_inplace_triton_kernel_training, test/inductor/test_perf.py::InplacingTests::test_inplace_triton_kernel_v1, test/inductor/test_perf.py::InplacingTests::test_inplace_triton_kernel_v2, test/inductor/test_perf.py::InplacingTests::test_inplace_triton_kernel_v3, test/inductor/test_perf.py::InplacingTests::test_inplace_triton_kernel_v4, test/inductor/test_perf.py::InplacingTests::test_inplace_triton_kernel_v5, test/inductor/test_perf.py::InplacingTests::test_inplace_triton_kernel_v6, test/inductor/test_perf.py::InplacingTests::test_triton_kernel_not_fusable_with_users 2025-07-17T10:18:19.9891133Z 2025-07-17T10:18:19.9891272Z Running inductor/test_profiler 1/1 ... [2025-07-17 10:18:19.986329] 2025-07-17T10:18:19.9891613Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:18:19.9892275Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_profiler.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:18:19.986650] 2025-07-17T10:18:34.1832630Z 2025-07-17T10:18:34.1833469Z inductor/test_profiler 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_profiler_1.1_f72c4b82926e32d0_.log 2025-07-17T10:18:34.1836214Z Running 7 items in this shard: test/inductor/test_profiler.py::DynamoProfilerTests::test_cupti_lazy_reinit, test/inductor/test_profiler.py::DynamoProfilerTests::test_inductor_profiling_kernel_names_foreach, test/inductor/test_profiler.py::DynamoProfilerTests::test_inductor_profiling_kernel_names_pointwise, test/inductor/test_profiler.py::DynamoProfilerTests::test_inductor_profiling_kernel_names_template, test/inductor/test_profiler.py::DynamoProfilerTests::test_inductor_profiling_triton_hooks, test/inductor/test_profiler.py::DynamoProfilerTests::test_inductor_profiling_triton_launch, test/inductor/test_profiler.py::DynamoProfilerTests::test_pt2_triton_attributes 2025-07-17T10:18:34.1837879Z 2025-07-17T10:18:34.1839381Z Running inductor/test_quantization 1/1 ... [2025-07-17 10:18:34.183148] 2025-07-17T10:18:34.1839785Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:18:34.1840529Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_quantization.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:18:34.183494] 2025-07-17T10:19:00.3998170Z 2025-07-17T10:19:00.3999045Z inductor/test_quantization 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_quantization_1.1_caf9656512a53da6_.log 2025-07-17T10:19:00.4000058Z Running 2 items in this shard: test/inductor/test_quantization.py::TestQuantization::test_activation_quantization_aten_with_scaling, test/inductor/test_quantization.py::TestQuantization::test_activation_quantization_aten_without_scaling 2025-07-17T10:19:00.4000678Z 2025-07-17T10:19:00.4000859Z Running inductor/test_scatter_optimization 1/1 ... [2025-07-17 10:19:00.399731] 2025-07-17T10:19:00.4001180Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:19:00.4002212Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_scatter_optimization.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:19:00.400074] 2025-07-17T10:19:28.5094648Z 2025-07-17T10:19:28.5095563Z inductor/test_scatter_optimization 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_scatter_optimization_1.1_607a36de7f77a8cd_.log 2025-07-17T10:19:28.5098991Z Running 8 items in this shard: test/inductor/test_scatter_optimization.py::TestScatterOpt::test_3d_tensor, test/inductor/test_scatter_optimization.py::TestScatterOpt::test_can_not_optimize_due_to_dense, test/inductor/test_scatter_optimization.py::TestScatterOpt::test_can_not_optimize_due_to_non_const, test/inductor/test_scatter_optimization.py::TestScatterOpt::test_cross_entropy_loss, test/inductor/test_scatter_optimization.py::TestScatterOpt::test_neg_scatter_dim, test/inductor/test_scatter_optimization.py::TestScatterOpt::test_non_last_dim, test/inductor/test_scatter_optimization.py::TestScatterOpt::test_nonzero_const_tensor, test/inductor/test_scatter_optimization.py::TestScatterOpt::test_shorter_index_tensor 2025-07-17T10:19:28.5101375Z 2025-07-17T10:19:28.5103603Z Running inductor/test_select_algorithm 1/1 ... [2025-07-17 10:19:28.509410] 2025-07-17T10:19:28.5104061Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:19:28.5105026Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_select_algorithm.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:19:28.509755] 2025-07-17T10:20:16.3244845Z 2025-07-17T10:20:16.3245587Z inductor/test_select_algorithm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_select_algorithm_1.1_15eb6a99498647e5_.log 2025-07-17T10:20:16.3250382Z Running 20 items in this shard: test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_TritonTemplateCaller_str, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test__int_mm, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_addmm, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_addmm_fp16, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_baddbmm, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_bmm, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_convolution1, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_convolution2, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_convolution_as_mm, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_linear_relu, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_mm, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_mm_dropout, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_mm_dup_args, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_mm_dup_args_view, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_mm_not_even_k, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_mm_plus_mm, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_mm_plus_mm2, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_mm_plus_mm3, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_mm_skip, test/inductor/test_select_algorithm.py::TestSelectAlgorithm::test_preprocessing_single_choice 2025-07-17T10:20:16.3254311Z 2025-07-17T10:20:16.3254455Z Running inductor/test_snode_runtime 1/1 ... [2025-07-17 10:20:16.324428] 2025-07-17T10:20:16.3254736Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:20:16.3255409Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_snode_runtime.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:20:16.324742] 2025-07-17T10:20:35.1294017Z 2025-07-17T10:20:35.1294774Z inductor/test_snode_runtime 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_snode_runtime_1.1_84d622157c513352_.log 2025-07-17T10:20:35.1299817Z Running 22 items in this shard: test/inductor/test_snode_runtime.py::UnsupportedTests::test_no_cuda, test/inductor/test_snode_runtime.py::UnsupportedTests::test_no_op, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_addmm, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_bmm, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_conv1d, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_conv2d, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_conv2d_transpose, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_conv3d, test/inductor/test_snode_runtime.py::ComputeBoundedTests::test_mm, test/inductor/test_snode_runtime.py::MemoryBoundedTests::test_dynamic, test/inductor/test_snode_runtime.py::MemoryBoundedTests::test_horizontal_reduction_pointwise, test/inductor/test_snode_runtime.py::MemoryBoundedTests::test_pointwise, test/inductor/test_snode_runtime.py::MemoryBoundedTests::test_relu, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_all_gather_into_tensor, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_all_gather_into_tensor_coalesced, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_all_reduce, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_all_reduce_coalesced, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_legacy_all_gather_into_tensor_coalesced, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_legacy_all_reduce, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_legacy_all_reduce_coalesced, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_reduce_scatter_tensor, test/inductor/test_snode_runtime.py::TestCommAnalysis::test_reduce_scatter_tensor_coalesced 2025-07-17T10:20:35.1304189Z 2025-07-17T10:20:35.1304374Z Running inductor/test_split_cat_fx_aten_passes 1/1 ... [2025-07-17 10:20:35.129383] 2025-07-17T10:20:35.1304677Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:20:35.1305367Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_split_cat_fx_aten_passes.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:20:35.129709] 2025-07-17T10:20:48.5714571Z 2025-07-17T10:20:48.5715566Z inductor/test_split_cat_fx_aten_passes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_split_cat_fx_aten_passes_1.1_99de13d92d16f199_.log 2025-07-17T10:20:48.5717276Z Running 5 items in this shard: test/inductor/test_split_cat_fx_aten_passes.py::TestSplitCatAten::test_move_view_after_cat_aten, test/inductor/test_split_cat_fx_aten_passes.py::TestSplitCatAten::test_select_cat_post_grad, test/inductor/test_split_cat_fx_aten_passes.py::TestSplitCatAten::test_split_cat_post_grad, test/inductor/test_split_cat_fx_aten_passes.py::TestSplitCatAten::test_split_cat_post_grad_singular, test/inductor/test_split_cat_fx_aten_passes.py::TestSplitCatAtenNormalizationPasses::test_split_aten_normalization 2025-07-17T10:20:48.5718635Z 2025-07-17T10:20:48.5718807Z Running inductor/test_split_cat_fx_passes 1/1 ... [2025-07-17 10:20:48.571351] 2025-07-17T10:20:48.5719102Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:20:48.5719780Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_split_cat_fx_passes.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:20:48.571681] 2025-07-17T10:21:41.9478468Z 2025-07-17T10:21:41.9482070Z inductor/test_split_cat_fx_passes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_split_cat_fx_passes_1.1_581a77ba86c7e6ed_.log 2025-07-17T10:21:41.9485645Z Running 11 items in this shard: test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_cat_normalization, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_config_flag_is_respected, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_consecutive_split_merge, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_numpy_compat_normalization, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_split_cat_merge, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_split_cat_merge_mutation, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_split_cat_new_patterns, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_split_normalization, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_split_squeeze, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_stack_normalization_axis_kwarg, test/inductor/test_split_cat_fx_passes.py::TestSplitCatFxPasses::test_unbind_stack 2025-07-17T10:21:41.9488518Z 2025-07-17T10:21:41.9488691Z Running inductor/test_static_cuda_launcher 1/1 ... [2025-07-17 10:21:41.947736] 2025-07-17T10:21:41.9488987Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:21:41.9489922Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_static_cuda_launcher.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:21:41.948081] 2025-07-17T10:21:49.5804164Z 2025-07-17T10:21:49.5805003Z inductor/test_static_cuda_launcher 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_static_cuda_launcher_1.1_509a2faa0d8887c5_.log 2025-07-17T10:21:49.5810015Z Running 18 items in this shard: test/inductor/test_static_cuda_launcher.py::TestStaticCudaLauncher::test_basic, test/inductor/test_static_cuda_launcher.py::TestStaticCudaLauncher::test_basic_1arg, test/inductor/test_static_cuda_launcher.py::TestStaticCudaLauncher::test_constexpr, test/inductor/test_static_cuda_launcher.py::TestStaticCudaLauncher::test_floats, test/inductor/test_static_cuda_launcher.py::TestStaticCudaLauncher::test_high_shared_mem, test/inductor/test_static_cuda_launcher.py::TestStaticCudaLauncher::test_implied_constant, test/inductor/test_static_cuda_launcher.py::TestStaticCudaLauncher::test_kernel_empty_tensor, test/inductor/test_static_cuda_launcher.py::TestStaticCudaLauncher::test_kernel_many_args, test/inductor/test_static_cuda_launcher.py::TestStaticCudaLauncher::test_kernel_no_args, test/inductor/test_static_cuda_launcher.py::TestStaticCudaLauncher::test_signed_integers, test/inductor/test_static_cuda_launcher.py::TestStaticCudaLauncher::test_too_high_shared_mem, test/inductor/test_static_cuda_launcher.py::TestStaticCudaLauncher::test_unsigned_integers, test/inductor/test_static_cuda_launcher.py::TestStaticTritonCompileResult::test_any, test/inductor/test_static_cuda_launcher.py::TestStaticTritonCompileResult::test_basic_compile, test/inductor/test_static_cuda_launcher.py::TestStaticTritonCompileResult::test_disable_static_cuda_launcher, test/inductor/test_static_cuda_launcher.py::TestStaticTritonCompileResult::test_empty_tensor, test/inductor/test_static_cuda_launcher.py::TestStaticTritonCompileResult::test_incompatible_code, test/inductor/test_static_cuda_launcher.py::TestStaticTritonCompileResult::test_static_launch_user_defined_triton_kernels 2025-07-17T10:21:49.5814187Z 2025-07-17T10:21:49.5814350Z Running inductor/test_subgraph_choice 1/1 ... [2025-07-17 10:21:49.580415] 2025-07-17T10:21:49.5814639Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:21:49.5815314Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_subgraph_choice.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:21:49.580731] 2025-07-17T10:21:57.4627546Z 2025-07-17T10:21:57.4628373Z inductor/test_subgraph_choice 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_subgraph_choice_1.1_b27e3439ea118fb8_.log 2025-07-17T10:21:57.4629357Z Running 2 items in this shard: test/inductor/test_subgraph_choice.py::TestSubgraphChoice::test_subgraph_decompose_k, test/inductor/test_subgraph_choice.py::TestSubgraphChoice::test_subgraph_freeze_layout 2025-07-17T10:21:57.4629873Z 2025-07-17T10:21:57.4630034Z Running inductor/test_torchbind 1/1 ... [2025-07-17 10:21:57.462657] 2025-07-17T10:21:57.4630303Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:21:57.4631716Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchbind.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:21:57.462974] 2025-07-17T10:22:36.3839961Z 2025-07-17T10:22:36.3843549Z inductor/test_torchbind 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchbind_1.1_d8e23afb4784bfb0_.log 2025-07-17T10:22:36.3847614Z Running 16 items in this shard: test/inductor/test_torchbind.py::TestTorchbind::test_aoti_torchbind_name_collision, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_aot_compile, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_aot_compile_constant_folding, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_aoti, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_compile, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_compile_gpu_op_symint_graph_partition, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_compile_symint, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_config_not_generated, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_get_buf_bytes, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_hop_schema, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_hop_schema_no_input, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_hop_schema_no_output, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_inductor, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_input_aot_compile, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_list_return_aot_compile, test/inductor/test_torchbind.py::TestTorchbind::test_torchbind_queue 2025-07-17T10:22:36.3850948Z 2025-07-17T10:22:36.3851170Z Running inductor/test_torchinductor_codegen_config_overrides 1/1 ... [2025-07-17 10:22:36.383987] 2025-07-17T10:22:36.3851514Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:22:36.3852249Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor_codegen_config_overrides.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:22:36.384322] 2025-07-17T10:22:51.2319650Z 2025-07-17T10:22:51.2320535Z inductor/test_torchinductor_codegen_config_overrides 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_codegen_config_overrides_1.1_8449af0eec916e69_.log 2025-07-17T10:22:51.2322140Z Running 3 items in this shard: test/inductor/test_torchinductor_codegen_config_overrides.py::CodegenInductorTest::test_force_pointwise_cat_force_pointwise_cat_False, test/inductor/test_torchinductor_codegen_config_overrides.py::CodegenInductorTest::test_force_pointwise_cat_force_pointwise_cat_True, test/inductor/test_torchinductor_codegen_config_overrides.py::CodegenInductorTest::test_kernel_fusion_thresholds 2025-07-17T10:22:51.2323162Z 2025-07-17T10:22:51.2323346Z Running inductor/test_torchinductor_opinfo 1/13 ... [2025-07-17 10:22:51.231858] 2025-07-17T10:22:51.2323654Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:22:51.2324357Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'not serial', '--shard-id=1', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:22:51.232204] 2025-07-17T10:23:57.5996981Z 2025-07-17T10:23:57.5997721Z inductor/test_flex_decoding 1/3 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_flex_decoding_1.3_40cd4036ecc77a84_.log 2025-07-17T10:23:57.6077152Z Running 210 items in this shard: test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_bfloat16_score_mod1_head_dims1_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_bfloat16_score_mod2_head_dims2_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_bfloat16_score_mod3_head_dims1_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_bfloat16_score_mod4_head_dims0_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_bfloat16_score_mod4_head_dims2_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_bfloat16_score_mod5_head_dims0_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_bfloat16_score_mod5_head_dims2_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_bfloat16_score_mod6_head_dims1_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_bfloat16_score_mod7_head_dims2_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_bfloat16_score_mod8_head_dims0_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_bfloat16_score_mod0_BLOCK_SIZE3_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_bfloat16_score_mod0_BLOCK_SIZE_128_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_bfloat16_score_mod1_BLOCK_SIZE2_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_bfloat16_score_mod1_BLOCK_SIZE_128_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_bfloat16_score_mod2_BLOCK_SIZE3_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_bfloat16_score_mod4_BLOCK_SIZE3_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_bfloat16_score_mod5_BLOCK_SIZE_64_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_bfloat16_score_mod6_BLOCK_SIZE3_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_bfloat16_score_mod6_BLOCK_SIZE_64_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_bfloat16_score_mod7_BLOCK_SIZE3_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_bfloat16_score_mod7_BLOCK_SIZE_64_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_bfloat16_score_mod8_BLOCK_SIZE_64_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod0_BLOCK_SIZE_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod0_BLOCK_SIZE_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod1_BLOCK_SIZE2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod1_BLOCK_SIZE3_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod2_BLOCK_SIZE_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod3_BLOCK_SIZE2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod3_BLOCK_SIZE_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod4_BLOCK_SIZE2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod4_BLOCK_SIZE3_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod4_BLOCK_SIZE_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod5_BLOCK_SIZE2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod5_BLOCK_SIZE3_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod5_BLOCK_SIZE_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod5_BLOCK_SIZE_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float16_score_mod7_BLOCK_SIZE2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod0_BLOCK_SIZE2_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod0_BLOCK_SIZE_128_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod1_BLOCK_SIZE3_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod2_BLOCK_SIZE3_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod2_BLOCK_SIZE_128_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod2_BLOCK_SIZE_64_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod3_BLOCK_SIZE_64_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod4_BLOCK_SIZE2_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod4_BLOCK_SIZE_128_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod5_BLOCK_SIZE2_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod5_BLOCK_SIZE_128_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod6_BLOCK_SIZE_128_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod7_BLOCK_SIZE3_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod7_BLOCK_SIZE_128_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod8_BLOCK_SIZE2_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_different_block_size_float32_score_mod8_BLOCK_SIZE_128_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float16_score_mod0_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float16_score_mod2_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float16_score_mod2_head_dims2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float16_score_mod4_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float16_score_mod4_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float16_score_mod6_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float16_score_mod7_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float16_score_mod8_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float16_score_mod8_head_dims2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float32_score_mod1_head_dims2_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float32_score_mod2_head_dims0_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float32_score_mod4_head_dims1_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float32_score_mod4_head_dims2_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float32_score_mod5_head_dims0_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float32_score_mod5_head_dims1_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float32_score_mod5_head_dims2_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float32_score_mod6_head_dims0_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_builtin_score_mods_float32_score_mod8_head_dims2_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_bw_decoding_fails_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_captured_buffers_all_dims_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_captured_buffers_all_dims_float32_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_captured_buffers_bfloat16_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_captured_buffers_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_decode_at_different_input_position_float16_score_mod2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_decode_at_different_input_position_float16_score_mod3_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_decode_at_different_input_position_float16_score_mod6_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_decode_at_different_input_position_float16_score_mod7_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_fully_masked_out_rows_0_check_gqa_cuda, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_function_composition_bfloat16_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_function_composition_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_head_dependent_mask_mod_float16_score_mod0_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_head_dependent_mask_mod_float16_score_mod3_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_head_dependent_mask_mod_float16_score_mod4_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_head_dependent_mask_mod_float16_score_mod4_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_head_dependent_mask_mod_float16_score_mod5_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_head_dependent_mask_mod_float16_score_mod6_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_head_dependent_mask_mod_float16_score_mod6_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_head_dependent_mask_mod_float16_score_mod8_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims0_score_mod5_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims0_score_mod7_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims0_score_mod8_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims1_score_mod3_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims1_score_mod5_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims2_score_mod0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims2_score_mod2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims3_score_mod1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims3_score_mod2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims3_score_mod3_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims3_score_mod4_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims3_score_mod5_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims0_batch_dims3_score_mod7_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims1_batch_dims0_score_mod2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims1_batch_dims0_score_mod3_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims1_batch_dims1_score_mod1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims1_batch_dims1_score_mod7_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims1_batch_dims2_score_mod0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims1_batch_dims2_score_mod1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims1_batch_dims3_score_mod2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims1_batch_dims3_score_mod3_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims1_batch_dims3_score_mod7_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims0_score_mod0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims1_score_mod2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims1_score_mod3_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims1_score_mod4_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims1_score_mod8_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims2_score_mod3_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims2_score_mod4_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims2_score_mod5_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims2_score_mod7_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims3_score_mod4_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims3_score_mod5_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims3_score_mod6_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_kv_batch_broadcast_float16_head_dims2_batch_dims3_score_mod7_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_larger_block_mask_bug_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_logsumexp_correctness_float16_score_mod0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_logsumexp_correctness_float32_score_mod1_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_logsumexp_only_return_cuda, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_multiple_score_mod_calls2_cuda, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_multiple_score_mod_calls_cuda, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_njt_causal_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_divisible_multi_token_offset_mask_cuda, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod0_float16_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod0_float32_head_dims0_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod0_float32_head_dims1_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod1_float16_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod2_bfloat16_head_dims1_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod2_float16_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod2_float32_head_dims0_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod3_float32_head_dims0_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod4_float16_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod4_float32_head_dims1_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod5_bfloat16_head_dims1_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod6_float16_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod6_float32_head_dims0_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod6_float32_head_dims1_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod7_bfloat16_head_dims1_cuda_bfloat16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod7_float16_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod7_float32_head_dims0_cuda_float32, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_equal_head_dims_score_mod8_float16_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_pow_2_headdim_head_dim_121_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_pow_2_headdim_head_dim_17_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_pow_2_headdim_head_dim_24_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_pow_2_headdim_head_dim_94_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_non_sparse_mulitple_block_size_cuda, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod0_head_dims0_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod0_head_dims1_page_size_256_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod0_head_dims2_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod1_head_dims1_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod1_head_dims1_page_size_256_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod1_head_dims1_page_size_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod1_head_dims2_page_size_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod2_head_dims0_page_size_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod2_head_dims2_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod2_head_dims2_page_size_256_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod2_head_dims2_page_size_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod3_head_dims0_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod3_head_dims1_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod3_head_dims1_page_size_256_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod3_head_dims2_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod3_head_dims2_page_size_256_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod4_head_dims0_page_size_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod4_head_dims2_page_size_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod5_head_dims0_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod5_head_dims0_page_size_256_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod5_head_dims0_page_size_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod5_head_dims2_page_size_256_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod6_head_dims0_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod6_head_dims0_page_size_256_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod6_head_dims0_page_size_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod6_head_dims1_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod6_head_dims2_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod6_head_dims2_page_size_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod7_head_dims0_page_size_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod8_head_dims0_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod8_head_dims1_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod8_head_dims1_page_size_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod8_head_dims2_page_size_128_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_paged_attention_page_size_float16_score_mod8_head_dims2_page_size_64_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_recompile_changed_score_mod_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_seq_masking_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_silu_on_score_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_skip_odd_keys_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_strided_inputs_float16_k_s0_v_s3_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_strided_inputs_float16_k_s1_v_s1_head_dims2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_strided_inputs_float16_k_s1_v_s2_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_strided_inputs_float16_k_s1_v_s2_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_strided_inputs_float16_k_s1_v_s2_head_dims2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_strided_inputs_float16_k_s1_v_s3_head_dims2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_strided_inputs_float16_k_s2_v_s1_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_strided_inputs_float16_k_s2_v_s2_head_dims0_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_strided_inputs_float16_k_s2_v_s2_head_dims1_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_strided_inputs_float16_k_s2_v_s2_head_dims2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_strided_inputs_float16_k_s3_v_s1_head_dims2_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_subgraph_respect_decompostion_float16_cuda_float16, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_windowed_no_mask_vs_sdpa_cuda, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_windowed_no_mask_vs_sdpa_paged_attention_cuda, test/inductor/test_flex_decoding.py::TestFlexDecodingCUDA::test_windowed_partial_block_vs_sdpa_cuda 2025-07-17T10:23:57.6139836Z 2025-07-17T10:23:57.6140023Z Running inductor/test_torchinductor_opinfo 8/13 ... [2025-07-17 10:23:57.599837] 2025-07-17T10:23:57.6140337Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:23:57.6141036Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'not serial', '--shard-id=8', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:23:57.600134] 2025-07-17T10:31:24.6453238Z 2025-07-17T10:31:24.6454407Z inductor/test_torchinductor_opinfo 1/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_1.13_835648f326e29ed8_.log 2025-07-17T10:31:24.6526434Z Running 256 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_H_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rxor___cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__native_batch_norm_legit_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__unsafe_masked_index_put_accumulate_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_add_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_addcmul_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_all_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_amax_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_amin_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_aminmax_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_any_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_arange_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_arange_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_arange_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_argmax_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_argmin_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_argwhere_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_as_strided_copy_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_as_strided_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_asin_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atleast_2d_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bfloat16_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cat_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cdist_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ceil_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_char_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_clamp_min_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_column_stack_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_column_stack_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_complex_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_contiguous_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_copysign_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cummin_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_deg2rad_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_deg2rad_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diag_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diagonal_copy_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diagonal_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_digamma_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_div_no_rounding_mode_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_einsum_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_empty_strided_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_empty_strided_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_eq_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_equal_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_erf_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_exp_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_expand_as_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_expand_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_expm1_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_exponential_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fft2_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fftn_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_hfft2_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ifftshift_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_irfft2_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_rfft_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fill_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_flip_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_float_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fmax_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fmin_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_frac_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_full_like_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_gradient_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_fill_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_reduce_mean_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_select_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_select_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isclose_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isin_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isinf_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isnan_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isnan_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_item_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_item_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_2inputs_2outputs_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_4inputs_with_extra_args_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_binary_return_by_ref_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_kron_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_kthvalue_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_kthvalue_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_eigvalsh_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_lstsq_grad_oriented_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_norm_subgradients_at_zero_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_pinv_hermitian_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_solve_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_tensorsolve_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_vector_norm_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linspace_tensor_overload_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log1p_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log1p_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log2_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logdet_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logical_and_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logical_not_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logspace_tensor_overload_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mT_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_argmax_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_argmin_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_cumprod_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_cumsum_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_cumsum_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_fill_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_median_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_normalize_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_prod_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_scatter_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_softmax_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_sum_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_var_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_meshgrid_variadic_tensors_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_binary_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_reduction_no_dim_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_minimum_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_multinomial_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mvlgamma_mvlgamma_p_1_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nanmedian_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_narrow_copy_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_native_batch_norm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_neg_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_neg_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_empty_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_full_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_full_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_full_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_zeros_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_max_pool2d_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_alpha_dropout_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_binary_cross_entropy_with_logits_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_channel_shuffle_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_conv3d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_conv_transpose3d_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_dropout3d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_dropout3d_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_dropout3d_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_dropout_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_hardshrink_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_hardsigmoid_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_hardswish_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_hinge_embedding_loss_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bicubic_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_trilinear_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_trilinear_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_linear_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_local_response_norm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_margin_ranking_loss_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_multilabel_soft_margin_loss_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_constant_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_reflect_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_prelu_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_relu_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_rms_norm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_softmin_with_dtype_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_softplus_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_tanhshrink_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nonzero_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nonzero_static_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_norm_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_norm_fro_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_norm_inf_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_outer_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_outer_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_0_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_1_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_3_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_4_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_4_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_pow_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_prod_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_prod_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_randint_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ravel_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_real_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_real_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_real_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_remainder_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_renorm_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_renorm_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_repeat_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_repeat_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_repeat_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_repeat_interleave_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_reshape_as_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_reshape_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_resize_as__cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_resolve_conj_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_roll_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_roll_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_roll_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rot90_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_round_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_round_decimals_neg_3_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsqrt_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_add_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_select_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_select_scatter_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sigmoid_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sign_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_signal_windows_cosine_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_signal_windows_general_hamming_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_signal_windows_hann_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sinc_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sinc_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_slice_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_slice_scatter_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_softmax_with_dtype_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_softmax_with_dtype_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_softmax_with_dtype_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_bessel_j1_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_bessel_y0_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_chebyshev_polynomial_t_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_i1e_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_legendre_polynomial_p_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_modified_bessel_i0_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_scaled_modified_bessel_k1_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_v_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_w_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_zeta_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_with_sizes_copy_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_square_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_squeeze_copy_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_squeeze_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_squeeze_multiple_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stack_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_std_mean_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sum_to_size_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sum_to_size_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_svd_lowrank_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_take_along_dim_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tile_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tile_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_to_sparse_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trace_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trapezoid_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trapz_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tril_indices_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_triu_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_true_divide_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unflatten_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unfold_copy_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_uniform_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unravel_index_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unsafe_split_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_view_copy_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_vsplit_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_where_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_xlogy_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_zeros_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_zeros_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_zeros_like_cuda_float32 2025-07-17T10:31:24.6597194Z 2025-07-17T10:31:24.6597375Z Running inductor/test_torchinductor_opinfo 11/13 ... [2025-07-17 10:31:24.645708] 2025-07-17T10:31:24.6597691Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:31:24.6598395Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '-m', 'not serial', '--shard-id=11', '--num-shards=13', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:31:24.646035] 2025-07-17T10:32:44.9839770Z 2025-07-17T10:32:44.9840772Z inductor/test_torchinductor_opinfo 8/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_8.13_06569ab19c60c478_.log 2025-07-17T10:32:44.9916272Z Running 269 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_H_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___getitem___cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rand___cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rmul___cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rsub___cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rxor___cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__segment_reduce_offsets_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_abs_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_acosh_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_addcmul_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_addmv_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_argsort_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_argwhere_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_as_strided_partial_views_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_asin_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atan2_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atan2_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atan2_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atan_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atan_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atleast_2d_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bitwise_and_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bitwise_or_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bmm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bool_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_broadcast_tensors_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bucketize_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_byte_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cartesian_prod_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cat_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cat_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cdouble_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cdouble_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cfloat_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cfloat_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cholesky_inverse_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_chunk_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_clamp_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_clamp_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_clone_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_contiguous_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_corrcoef_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cosh_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_count_nonzero_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cov_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cummax_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cummin_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diag_embed_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diagonal_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diagonal_scatter_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_dsplit_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_empty_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_empty_like_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_eq_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_erf_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_exp2_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_exp2_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_exponential_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fft_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fftn_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_hfft2_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_hfft2_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_hfft_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ifft2_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ifft_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ifftshift_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ifftshift_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ihfft2_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_irfft2_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_rfft_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_rfftn_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_rfftn_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_float_power_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_floor_divide_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fmax_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_frexp_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_full_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_full_like_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_gcd_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ge_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ge_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_geometric_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_gt_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_hsplit_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_hsplit_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_hsplit_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_hstack_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_add_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_add_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isnan_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isnan_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isposinf_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_2inputs_2outputs_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_unary_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_kron_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ldexp_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_det_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_ldl_solve_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_lstsq_grad_oriented_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_matrix_power_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_matrix_rank_hermitian_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_multi_dot_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_pinv_hermitian_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_vander_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_vecdot_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log10_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log1p_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logical_and_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logical_xor_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logical_xor_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logspace_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logspace_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logspace_tensor_overload_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_amin_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_argmax_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_argmin_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_select_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_select_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_softmax_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matmul_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_max_binary_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_max_reduction_no_dim_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_maximum_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_median_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_binary_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_binary_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_reduction_no_dim_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_reduction_no_dim_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_reduction_with_dim_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_minimum_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mm_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mv_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mvlgamma_mvlgamma_p_5_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_narrow_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool1d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_binary_cross_entropy_with_logits_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_channel_shuffle_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_channel_shuffle_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_conv_transpose2d_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_dropout_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_embedding_bag_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_feature_alpha_dropout_without_train_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_gelu_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_gelu_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_hardshrink_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_hardswish_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bicubic_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_nearest_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_logsigmoid_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_max_pool1d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_max_pool2d_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_max_unpool1d_grad_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_max_unpool2d_grad_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_max_unpool3d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_multi_margin_loss_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_replicate_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pairwise_distance_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pixel_shuffle_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pixel_unshuffle_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pixel_unshuffle_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_poisson_nll_loss_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_poisson_nll_loss_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_poisson_nll_loss_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_prelu_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_relu6_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_relu6_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_relu_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_rms_norm_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_rrelu_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_scaled_dot_product_attention_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_threshold_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_threshold_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_triplet_margin_with_distance_loss_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nonzero_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nonzero_static_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nonzero_static_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_norm_fro_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_normal_in_place_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ones_like_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_outer_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_permute_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_permute_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_0_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_0_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_1_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_2_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_3_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_prod_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rand_like_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_randint_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_randint_like_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_randn_like_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_reciprocal_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_resize__cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_round_decimals_3_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rsub_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_reduce_amax_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_searchsorted_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_select_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_select_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_short_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_short_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sigmoid_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sinh_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_softmax_with_dtype_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sort_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_airy_ai_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_airy_ai_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_chebyshev_polynomial_t_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_chebyshev_polynomial_v_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_chebyshev_polynomial_w_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_hermite_polynomial_h_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_hermite_polynomial_he_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_hermite_polynomial_he_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_i0e_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_i1e_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_laguerre_polynomial_l_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_legendre_polynomial_p_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_modified_bessel_i0_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_modified_bessel_k1_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_ndtri_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_ndtri_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_polygamma_special_polygamma_n_0_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_scaled_modified_bessel_k0_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_scaled_modified_bessel_k1_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_w_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_xlog1py_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_list_args_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_with_sizes_copy_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_with_sizes_copy_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_squeeze_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_squeeze_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_std_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sub_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_t_copy_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_t_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_take_along_dim_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tan_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tile_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_to_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_transpose_copy_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_transpose_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trapz_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_triu_indices_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_true_divide_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trunc_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unfold_copy_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unique_consecutive_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unique_consecutive_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unique_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unsqueeze_copy_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unsqueeze_copy_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_var_unbiased_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_vdot_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_view_as_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_view_copy_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_view_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_vsplit_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_vstack_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_vstack_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_where_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_xlogy_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_zero__cuda_bool 2025-07-17T10:32:44.9990454Z 2025-07-17T10:32:44.9990595Z Running optim/test_swa_utils 1/1 ... [2025-07-17 10:32:44.984283] 2025-07-17T10:32:44.9990875Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:32:44.9991528Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'optim/test_swa_utils.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:32:44.984608] 2025-07-17T10:32:48.3133591Z 2025-07-17T10:32:48.3135258Z optim/test_swa_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/optim.test_swa_utils_1.1_562b970e100d9a49_.log 2025-07-17T10:32:48.3135792Z 2025-07-17T10:32:48.3135953Z Running test_out_dtype_op 1/1 ... [2025-07-17 10:32:48.313222] 2025-07-17T10:32:48.3136228Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:32:48.3137048Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_out_dtype_op.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:32:48.313537] 2025-07-17T10:32:56.8962483Z 2025-07-17T10:32:56.8963569Z test_out_dtype_op 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_out_dtype_op_1.1_902a6948009dd6c1_.log 2025-07-17T10:32:56.8966596Z Running 12 items in this shard: test/test_out_dtype_op.py::TestOutDtypeOp::test_out_dtype_dynamo, test/test_out_dtype_op.py::TestOutDtypeOp::test_out_dtype_inductor_decomp, test/test_out_dtype_op.py::TestOutDtypeOp::test_out_dtype_inductor_decomp_trace, test/test_out_dtype_op.py::TestOutDtypeOp::test_out_dtype_int_mm_default_trace, test/test_out_dtype_op.py::TestOutDtypeOp::test_out_dtype_make_fx, test/test_out_dtype_op.py::TestOutDtypeOp::test_out_dtype_mm_numerical, test/test_out_dtype_op.py::TestOutDtypeOp::test_out_dtype_mul_scalar_numerical, test/test_out_dtype_op.py::TestOutDtypeOp::test_out_dtype_no_autograd, test/test_out_dtype_op.py::TestOutDtypeOp::test_out_dtype_non_functional, test/test_out_dtype_op.py::TestOutDtypeOp::test_out_dtype_non_op_overload, test/test_out_dtype_op.py::TestOutDtypeOp::test_out_dtype_op_functional, test/test_out_dtype_op.py::TestOutDtypeOp::test_out_dtype_wrong_output 2025-07-17T10:32:56.8969727Z 2025-07-17T10:32:56.8969882Z Running test_package 1/1 ... [2025-07-17 10:32:56.896273] 2025-07-17T10:32:56.8970190Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:32:56.8970995Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_package.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:32:56.896649] 2025-07-17T10:33:02.0220343Z 2025-07-17T10:33:02.0221309Z test_package 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_package_1.1_711b0a2b024f64d8_.log 2025-07-17T10:33:02.0242439Z Running 137 items in this shard: test/test_package.py::TestAnalyze::test_trace_dependencies, test/test_package.py::TestDependencyAPI::test_allow_empty_with_error, test/test_package.py::TestDependencyAPI::test_broken_dependency, test/test_package.py::TestDependencyAPI::test_deny, test/test_package.py::TestDependencyAPI::test_deny_glob, test/test_package.py::TestDependencyAPI::test_extern, test/test_package.py::TestDependencyAPI::test_extern_glob, test/test_package.py::TestDependencyAPI::test_extern_glob_allow_empty, test/test_package.py::TestDependencyAPI::test_externing_c_extension, test/test_package.py::TestDependencyAPI::test_implicit_intern, test/test_package.py::TestDependencyAPI::test_intern_error, test/test_package.py::TestDependencyAPI::test_invalid_import, test/test_package.py::TestDependencyAPI::test_mock, test/test_package.py::TestDependencyAPI::test_mock_glob, test/test_package.py::TestDependencyAPI::test_mock_glob_allow_empty, test/test_package.py::TestDependencyAPI::test_pickle_mocked, test/test_package.py::TestDependencyAPI::test_pickle_mocked_all, test/test_package.py::TestDependencyAPI::test_repackage_mocked_module, test/test_package.py::TestDependencyHooks::test_extern_and_mock_hook, test/test_package.py::TestDependencyHooks::test_multiple_extern_hooks, test/test_package.py::TestDependencyHooks::test_multiple_mock_hooks, test/test_package.py::TestDependencyHooks::test_remove_hooks, test/test_package.py::TestDependencyHooks::test_single_hook, test/test_package.py::TestDiGraph::test_all_paths, test/test_package.py::TestDiGraph::test_contains, test/test_package.py::TestDiGraph::test_contains_non_hashable, test/test_package.py::TestDiGraph::test_edges, test/test_package.py::TestDiGraph::test_forward_closure, test/test_package.py::TestDiGraph::test_iter, test/test_package.py::TestDiGraph::test_node_attr_update, test/test_package.py::TestDiGraph::test_node_attrs, test/test_package.py::TestDiGraph::test_predecessor_not_in_graph, test/test_package.py::TestDiGraph::test_predecessors, test/test_package.py::TestDiGraph::test_successor_not_in_graph, test/test_package.py::TestDiGraph::test_successors, test/test_package.py::DirectoryReaderTest::test_importer_access, test/test_package.py::DirectoryReaderTest::test_loading_has_record, test/test_package.py::DirectoryReaderTest::test_loading_module, test/test_package.py::DirectoryReaderTest::test_loading_pickle, test/test_package.py::DirectoryReaderTest::test_package_resource_access, test/test_package.py::DirectoryReaderTest::test_resource_access_by_path, test/test_package.py::DirectoryReaderTest::test_resource_reader, test/test_package.py::DirectoryReaderTest::test_scriptobject_failure_message, test/test_package.py::TestGlobGroup::test_exclude, test/test_package.py::TestGlobGroup::test_exclude_from_all, test/test_package.py::TestGlobGroup::test_invalid_raw, test/test_package.py::TestGlobGroup::test_list_include_exclude, test/test_package.py::TestGlobGroup::test_one_star, test/test_package.py::TestGlobGroup::test_one_star_middle, test/test_package.py::TestGlobGroup::test_one_star_multiple_in_component, test/test_package.py::TestGlobGroup::test_one_star_partial, test/test_package.py::TestGlobGroup::test_one_star_partial_extension, test/test_package.py::TestGlobGroup::test_raw_two_star, test/test_package.py::TestGlobGroup::test_two_star, test/test_package.py::TestGlobGroup::test_two_star_end, test/test_package.py::TestGlobGroup::test_two_star_middle, test/test_package.py::TestGlobGroup::test_two_star_multiple, test/test_package.py::TestImporter::test_ordered_importer_basic, test/test_package.py::TestImporter::test_ordered_importer_whichmodule, test/test_package.py::TestImporter::test_package_importer_whichmodule_no_dunder_module, test/test_package.py::TestImporter::test_single_ordered_importer, test/test_package.py::TestImporter::test_sys_importer, test/test_package.py::TestImporter::test_sys_importer_roundtrip, test/test_package.py::TestLoadBCPackages::test_load_bc_packages_fx_module, test/test_package.py::TestLoadBCPackages::test_load_bc_packages_nn_module, test/test_package.py::TestLoadBCPackages::test_load_bc_packages_torchscript_module, test/test_package.py::TestMangling::test_demangle_base, test/test_package.py::TestMangling::test_demangler_multiple_manglers, test/test_package.py::TestMangling::test_is_mangled, test/test_package.py::TestMangling::test_mangle_empty_errors, test/test_package.py::TestMangling::test_mangle_prefix, test/test_package.py::TestMangling::test_mangler_is_consistent, test/test_package.py::TestMangling::test_package_mangler, test/test_package.py::TestMangling::test_roundtrip_mangling, test/test_package.py::TestMangling::test_unique_manglers, test/test_package.py::TestMangling::test_unique_module_names, test/test_package.py::TestMisc::test_dunder_package_present, test/test_package.py::TestMisc::test_dunder_package_works_from_package, test/test_package.py::TestMisc::test_exporter_content_lists, test/test_package.py::TestMisc::test_file_structure, test/test_package.py::TestMisc::test_file_structure_has_file, test/test_package.py::TestMisc::test_inspect_class, test/test_package.py::TestMisc::test_is_from_package, test/test_package.py::TestMisc::test_load_python_version_from_package, test/test_package.py::TestMisc::test_loaders_that_remap_files_work_ok, test/test_package.py::TestMisc::test_python_version, test/test_package.py::TestMisc::test_std_lib_sys_hackery_checks, test/test_package.py::ModelTest::test_model_save, test/test_package.py::ModelTest::test_resnet, test/test_package.py::ModelTest::test_script_resnet, test/test_package.py::TestPackageFX::test_package_fx_custom_tracer, test/test_package.py::TestPackageFX::test_package_fx_package, test/test_package.py::TestPackageFX::test_package_fx_simple, test/test_package.py::TestPackageFX::test_package_fx_with_imports, test/test_package.py::TestPackageFX::test_package_fx_wrap, test/test_package.py::TestPackageFX::test_package_gm_preserve_stack_trace, test/test_package.py::TestPackageFX::test_package_then_fx, test/test_package.py::TestPackageScript::test_different_package_interface, test/test_package.py::TestPackageScript::test_different_package_script_class, test/test_package.py::TestPackageScript::test_load_shared_scriptmodules, test/test_package.py::TestPackageScript::test_load_shared_tensors, test/test_package.py::TestPackageScript::test_load_shared_tensors_repackaged, test/test_package.py::TestPackageScript::test_mixing_packaged_and_inline_modules, test/test_package.py::TestPackageScript::test_mixing_packaged_and_inline_modules_shared_code, test/test_package.py::TestPackageScript::test_package_interface, test/test_package.py::TestPackageScript::test_package_script_class, test/test_package.py::TestPackageScript::test_package_script_class_referencing_self, test/test_package.py::TestPackageScript::test_save_eager_mods_sharing_scriptmodule, test/test_package.py::TestPackageScript::test_save_independent_scriptmodules, test/test_package.py::TestPackageScript::test_save_repeat_scriptmodules, test/test_package.py::TestPackageScript::test_save_scriptmodule, test/test_package.py::TestPackageScript::test_save_scriptmodule_file, test/test_package.py::TestPackageScript::test_save_scriptmodule_only_necessary_code, test/test_package.py::TestPackageScript::test_save_scriptmodule_with_submods, test/test_package.py::TestPackageScript::test_save_scriptmodules_in_container, test/test_package.py::TestPackageScript::test_save_scriptmodules_submod_redefinition, test/test_package.py::TestPackageScript::test_save_shared_tensors, test/test_package.py::TestPackageScript::test_saving_and_scripting_packaged_mod, test/test_package.py::TestPackageScript::test_scriptmodules_repeat_save, test/test_package.py::TestPackageScript::test_tensor_sharing_pickle, test/test_package.py::TestRepackage::test_repackage_import_indirectly_via_parent_module, test/test_package.py::TestResources::test_importer_access, test/test_package.py::TestResources::test_package_resource_access, test/test_package.py::TestResources::test_resource_access_by_path, test/test_package.py::TestResources::test_resource_reader, test/test_package.py::TestSaveLoad::test_bad_dunder_imports, test/test_package.py::TestSaveLoad::test_dunder_imports, test/test_package.py::TestSaveLoad::test_exporting_mismatched_code, test/test_package.py::TestSaveLoad::test_pickle, test/test_package.py::TestSaveLoad::test_pickle_long_name_with_protocol_4, test/test_package.py::TestSaveLoad::test_save_imported_module, test/test_package.py::TestSaveLoad::test_save_imported_module_using_package_importer, test/test_package.py::TestSaveLoad::test_save_load_fp8, test/test_package.py::TestSaveLoad::test_save_module, test/test_package.py::TestSaveLoad::test_save_module_binary, test/test_package.py::TestSaveLoad::test_saving_source, test/test_package.py::TestSaveLoad::test_saving_string 2025-07-17T10:33:02.0262549Z 2025-07-17T10:33:02.0262692Z Running test_per_overload_api 1/1 ... [2025-07-17 10:33:02.022069] 2025-07-17T10:33:02.0262968Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:33:02.0263635Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_per_overload_api.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:33:02.022372] 2025-07-17T10:33:05.6952671Z 2025-07-17T10:33:05.6953614Z test_per_overload_api 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_per_overload_api_1.1_0ad42cf501d9cfb8_.log 2025-07-17T10:33:05.6955541Z Running 3 items in this shard: test/test_per_overload_api.py::TestPerOverloadAPI::test_basics_opoverload, test/test_per_overload_api.py::TestPerOverloadAPI::test_basics_opoverloadpacket, test/test_per_overload_api.py::TestPerOverloadAPI::test_decompose 2025-07-17T10:33:05.6956325Z 2025-07-17T10:33:05.6956441Z Running test_prims 1/1 ... [2025-07-17 10:33:05.695119] 2025-07-17T10:33:05.6956693Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:33:05.6957451Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_prims.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:33:05.695465] 2025-07-17T10:33:11.1713710Z 2025-07-17T10:33:11.1714738Z test_prims 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_prims_1.1_ced10c71e7cb45d6_.log 2025-07-17T10:33:11.1718895Z Running 24 items in this shard: test/test_prims.py::TestPrimsBasic::test_check_deprecation_warning, test/test_prims.py::TestPrimsBasic::test_clone_complex, test/test_prims.py::TestPrimsBasic::test_mul_complex, test/test_prims.py::TestPrimsBasic::test_torch_ops, test/test_prims.py::TestPrimsCUDA::test_aten_overload_to_prims_cuda, test/test_prims.py::TestPrimsCUDA::test_broadcast_in_dim_cuda_float32, test/test_prims.py::TestPrimsCUDA::test_broadcast_in_dim_sum_cuda_float32, test/test_prims.py::TestPrimsCUDA::test_cbrt_prim_cuda_float64, test/test_prims.py::TestPrimsCUDA::test_cbrt_prim_cuda_int64, test/test_prims.py::TestPrimsCUDA::test_collapse_cuda_float32, test/test_prims.py::TestPrimsCUDA::test_functional_rng_wrappers_cuda_float32, test/test_prims.py::TestPrimsCUDA::test_memory_format_strides_cuda_float32, test/test_prims.py::TestPrimsCUDA::test_philox_rand_cuda_float32, test/test_prims.py::TestPrimsCUDA::test_reshape_view_method_cuda_float32, test/test_prims.py::TestPrimsCUDA::test_var_correction_0_cuda_float32, test/test_prims.py::TestPrimsCUDA::test_var_correction_1_cuda_float32, test/test_prims.py::TestRefsCUDA::test_constant_pad_nd_memory_format_cuda_float32, test/test_prims.py::TestRefsCUDA::test_inferred_tags_cuda, test/test_prims.py::TestRefsCUDA::test_infinite_loop_from_py_dispatcher_cuda, test/test_prims.py::TestRefsCUDA::test_linspace_with_complex_input_cuda, test/test_prims.py::TestRefsCUDA::test_logspace_with_complex_input_cuda, test/test_prims.py::TestRefsCUDA::test_unbind_cuda, test/test_prims.py::TestDecompCUDA::test_decomposition_method_vararg_ones_cuda_float32, test/test_prims.py::TestDecompCUDA::test_decomposition_method_vararg_permute_cuda_float32 2025-07-17T10:33:11.1723517Z 2025-07-17T10:33:11.1723660Z Running test_proxy_tensor 1/1 ... [2025-07-17 10:33:11.171299] 2025-07-17T10:33:11.1723932Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:33:11.1724581Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_proxy_tensor.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:33:11.171625] 2025-07-17T10:33:26.3718435Z 2025-07-17T10:33:26.3719386Z test_proxy_tensor 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_proxy_tensor_1.1_4d503837a5b5fefd_.log 2025-07-17T10:33:26.3752714Z Running 173 items in this shard: test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_allclose, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_amp_cache, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_constant_blowup, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_constant_proxy_tensor_mut, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_constant_random, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_constant_unbind, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_decomp_of_capture, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_decomposition_interpreter, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_empty_like_doesnt_burn_in_defaults, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_inplace_metadata, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_isolated_graphmodule, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_make_fx_model_double_param, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_make_fx_model_fwd_bwd, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_make_fx_model_fwd_bwd_wgtupdate, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_make_fx_overloads, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_make_fx_reentrant_dispatch, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_make_fx_simple, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_mode_tracing_factory_function, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_partial_decomp, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_pickle_issue89626, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_pr_86917, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_pre_dispatch_functionalization, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_pre_dispatch_functionalization_view_op, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_pre_dispatch_linear, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_pre_dispatch_mode_stack, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_pre_dispatch_no_grad, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_proxy_tensor, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_proxy_tensor_mode_with_decomp_table_preserves_proxy, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_resnet18_backward_trace, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_scalar_device, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_strides, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_tensor_constants, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_trace_subclasses, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_val_metadata_mutation, test/test_proxy_tensor.py::TestGenericProxyTensorReal::test_varargs, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_allclose, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_amp_cache, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_constant_blowup, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_constant_proxy_tensor_mut, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_constant_random, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_constant_unbind, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_decomp_of_capture, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_decomposition_interpreter, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_empty_like_doesnt_burn_in_defaults, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_inplace_metadata, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_isolated_graphmodule, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_make_fx_model_double_param, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_make_fx_model_fwd_bwd, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_make_fx_model_fwd_bwd_wgtupdate, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_make_fx_overloads, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_make_fx_reentrant_dispatch, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_make_fx_simple, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_mode_tracing_factory_function, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_partial_decomp, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_pickle_issue89626, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_pr_86917, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_pre_dispatch_functionalization, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_pre_dispatch_functionalization_view_op, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_pre_dispatch_linear, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_pre_dispatch_mode_stack, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_pre_dispatch_no_grad, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_proxy_tensor, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_proxy_tensor_mode_with_decomp_table_preserves_proxy, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_resnet18_backward_trace, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_scalar_device, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_strides, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_tensor_constants, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_trace_subclasses, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_val_metadata_mutation, test/test_proxy_tensor.py::TestGenericProxyTensorFake::test_varargs, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_allclose, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_amp_cache, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_constant_blowup, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_constant_proxy_tensor_mut, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_constant_random, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_constant_unbind, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_decomp_of_capture, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_decomposition_interpreter, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_empty_like_doesnt_burn_in_defaults, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_inplace_metadata, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_isolated_graphmodule, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_make_fx_model_double_param, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_make_fx_model_fwd_bwd, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_make_fx_model_fwd_bwd_wgtupdate, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_make_fx_overloads, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_make_fx_reentrant_dispatch, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_make_fx_simple, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_mode_tracing_factory_function, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_partial_decomp, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_pickle_issue89626, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_pr_86917, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_pre_dispatch_functionalization, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_pre_dispatch_functionalization_view_op, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_pre_dispatch_linear, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_pre_dispatch_mode_stack, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_pre_dispatch_no_grad, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_proxy_tensor, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_proxy_tensor_mode_with_decomp_table_preserves_proxy, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_resnet18_backward_trace, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_scalar_device, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_strides, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_tensor_constants, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_trace_subclasses, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_val_metadata_mutation, test/test_proxy_tensor.py::TestGenericProxyTensorSymbolic::test_varargs, test/test_proxy_tensor.py::TestRealProxyTensor::test_error_on_data_dependent_ops, test/test_proxy_tensor.py::TestFakeProxyTensor::test_alias, test/test_proxy_tensor.py::TestFakeProxyTensor::test_fake_tensor_mode, test/test_proxy_tensor.py::TestFakeProxyTensor::test_free_fake, test/test_proxy_tensor.py::TestFakeProxyTensor::test_fused_adam, test/test_proxy_tensor.py::TestFakeProxyTensor::test_issue82547, test/test_proxy_tensor.py::TestFakeProxyTensor::test_meta, test/test_proxy_tensor.py::TestFakeProxyTensor::test_use_fake_and_tensor, test/test_proxy_tensor.py::TestSymbolicTracing::test_adv_index_batch, test/test_proxy_tensor.py::TestSymbolicTracing::test_arange_unbacked_output_size, test/test_proxy_tensor.py::TestSymbolicTracing::test_binary_broadcast, test/test_proxy_tensor.py::TestSymbolicTracing::test_boolean_index, test/test_proxy_tensor.py::TestSymbolicTracing::test_broadcast_shapes, test/test_proxy_tensor.py::TestSymbolicTracing::test_cat, test/test_proxy_tensor.py::TestSymbolicTracing::test_constant_specialization, test/test_proxy_tensor.py::TestSymbolicTracing::test_cpu_scalar_cuda, test/test_proxy_tensor.py::TestSymbolicTracing::test_cumsum_unbacked, test/test_proxy_tensor.py::TestSymbolicTracing::test_debug_interpreter, test/test_proxy_tensor.py::TestSymbolicTracing::test_deduped_shape, test/test_proxy_tensor.py::TestSymbolicTracing::test_dynamic_pointwise_scalar, test/test_proxy_tensor.py::TestSymbolicTracing::test_elementwise_meta_with_sym_numbers, test/test_proxy_tensor.py::TestSymbolicTracing::test_expand, test/test_proxy_tensor.py::TestSymbolicTracing::test_fake_tensor_as_size, test/test_proxy_tensor.py::TestSymbolicTracing::test_guard_lowerbound_range_refinement, test/test_proxy_tensor.py::TestSymbolicTracing::test_guard_lowerbound_range_refinement_multivariate, test/test_proxy_tensor.py::TestSymbolicTracing::test_guard_upperbound_range_refinement, test/test_proxy_tensor.py::TestSymbolicTracing::test_guard_upperbound_range_refinement_multivariate, test/test_proxy_tensor.py::TestSymbolicTracing::test_guards_equal, test/test_proxy_tensor.py::TestSymbolicTracing::test_int_input, test/test_proxy_tensor.py::TestSymbolicTracing::test_invalidate_nonzero, test/test_proxy_tensor.py::TestSymbolicTracing::test_invalidate_nonzero_propagate_real_tensors, test/test_proxy_tensor.py::TestSymbolicTracing::test_item, test/test_proxy_tensor.py::TestSymbolicTracing::test_item_to_constructor, test/test_proxy_tensor.py::TestSymbolicTracing::test_make_fx_with_custom_tracer_preserving_nn_module_stack, test/test_proxy_tensor.py::TestSymbolicTracing::test_mega_guard, test/test_proxy_tensor.py::TestSymbolicTracing::test_metadata, test/test_proxy_tensor.py::TestSymbolicTracing::test_metadata_fresh, test/test_proxy_tensor.py::TestSymbolicTracing::test_mod_gcd_unbacked, test/test_proxy_tensor.py::TestSymbolicTracing::test_multiply_shape, test/test_proxy_tensor.py::TestSymbolicTracing::test_neg_shape, test/test_proxy_tensor.py::TestSymbolicTracing::test_new_empty, test/test_proxy_tensor.py::TestSymbolicTracing::test_non_deduped_shape, test/test_proxy_tensor.py::TestSymbolicTracing::test_non_symint_size_spec, test/test_proxy_tensor.py::TestSymbolicTracing::test_nonidentity_transitive_guards, test/test_proxy_tensor.py::TestSymbolicTracing::test_reflect_r_over_x, test/test_proxy_tensor.py::TestSymbolicTracing::test_repeat_interleave, test/test_proxy_tensor.py::TestSymbolicTracing::test_repeat_interleave_unbacked_output_size, test/test_proxy_tensor.py::TestSymbolicTracing::test_reshape_divisibility_unbacked, test/test_proxy_tensor.py::TestSymbolicTracing::test_resize_from_zero, test/test_proxy_tensor.py::TestSymbolicTracing::test_return_symint, test/test_proxy_tensor.py::TestSymbolicTracing::test_rmethod, test/test_proxy_tensor.py::TestSymbolicTracing::test_setitem_symint, test/test_proxy_tensor.py::TestSymbolicTracing::test_size_with_tensor, test/test_proxy_tensor.py::TestSymbolicTracing::test_split_unbacked_sizes, test/test_proxy_tensor.py::TestSymbolicTracing::test_sqrt_size, test/test_proxy_tensor.py::TestSymbolicTracing::test_sym_storage_offset, test/test_proxy_tensor.py::TestSymbolicTracing::test_symbolic_repeat_interleave, test/test_proxy_tensor.py::TestSymbolicTracing::test_symint_to_tensor, test/test_proxy_tensor.py::TestSymbolicTracing::test_tensor_symfloat, test/test_proxy_tensor.py::TestSymbolicTracing::test_unary, test/test_proxy_tensor.py::TestSymbolicTracing::test_unbacked_batch_resnet, test/test_proxy_tensor.py::TestSymbolicTracing::test_unbacked_slice, test/test_proxy_tensor.py::TestSymbolicTracing::test_unbacked_unification, test/test_proxy_tensor.py::TestSymbolicTracing::test_unbacked_unify_dependency_violation, test/test_proxy_tensor.py::TestSymbolicTracing::test_unbacked_unify_guard, test/test_proxy_tensor.py::TestSymbolicTracing::test_unbacked_unify_guard_transitivity, test/test_proxy_tensor.py::TestSymbolicTracing::test_view_divisibility_unbacked, test/test_proxy_tensor.py::TestSymbolicTracing::test_view_divisibility_unbacked_relatively_prime 2025-07-17T10:33:26.3785207Z 2025-07-17T10:33:26.3785345Z Running test_pruning_op 1/1 ... [2025-07-17 10:33:26.371921] 2025-07-17T10:33:26.3785616Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:33:26.3786277Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_pruning_op.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:33:26.372251] 2025-07-17T10:33:30.1962387Z 2025-07-17T10:33:30.1963385Z test_pruning_op 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_pruning_op_1.1_8bc24d0236161020_.log 2025-07-17T10:33:30.1964258Z Running 2 items in this shard: test/test_pruning_op.py::PruningOpTest::test_rowwise_prune_op_32bit_indices, test/test_pruning_op.py::PruningOpTest::test_rowwise_prune_op_64bit_indices 2025-07-17T10:33:30.1964710Z 2025-07-17T10:33:30.1964866Z Running test_public_bindings 1/1 ... [2025-07-17 10:33:30.196176] 2025-07-17T10:33:30.1965131Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:33:30.1966455Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_public_bindings.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:33:30.196508] 2025-07-17T10:33:41.4916100Z 2025-07-17T10:33:41.4917066Z test_public_bindings 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_public_bindings_1.1_5ae7b283cd876fd8_.log 2025-07-17T10:33:41.4918327Z Running 4 items in this shard: test/test_public_bindings.py::TestPublicBindings::test_correct_module_names, test/test_public_bindings.py::TestPublicBindings::test_modules_can_be_imported, test/test_public_bindings.py::TestPublicBindings::test_no_new_bindings, test/test_public_bindings.py::TestPublicBindings::test_no_new_reexport_callables 2025-07-17T10:33:41.4919199Z 2025-07-17T10:33:41.4919329Z Running test_pytree 1/1 ... [2025-07-17 10:33:41.491416] 2025-07-17T10:33:41.4919579Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:33:41.4920228Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_pytree.py', '-m', 'not serial', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:33:41.491742] 2025-07-17T10:33:46.9679899Z 2025-07-17T10:33:46.9680817Z test_pytree 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_pytree_1.1_0cce168342610e18_.log 2025-07-17T10:33:46.9697292Z Running 98 items in this shard: test/test_pytree.py::TestGenericPytree::test_aligned_public_apis, test/test_pytree.py::TestGenericPytree::test_broadcast_to_and_flatten_cxx, test/test_pytree.py::TestGenericPytree::test_broadcast_to_and_flatten_py, test/test_pytree.py::TestGenericPytree::test_enum_treespec_roundtrip_cxx, test/test_pytree.py::TestGenericPytree::test_enum_treespec_roundtrip_py, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_defaultdict_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_defaultdict_py, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_deque_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_deque_py, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_dict_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_dict_py, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_leaf_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_leaf_py, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_list_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_list_py, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_namedtuple_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_namedtuple_py, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_nested_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_nested_py, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_ordereddict_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_ordereddict_py, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_return_types_max_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_return_types_max_py, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_return_types_min_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_return_types_min_py, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_tuple_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_unflatten_tuple_py, test/test_pytree.py::TestGenericPytree::test_flatten_with_is_leaf_cxx, test/test_pytree.py::TestGenericPytree::test_flatten_with_is_leaf_py, test/test_pytree.py::TestGenericPytree::test_is_namedtuple_cxx, test/test_pytree.py::TestGenericPytree::test_is_namedtuple_py, test/test_pytree.py::TestGenericPytree::test_is_structseq_cxx, test/test_pytree.py::TestGenericPytree::test_is_structseq_py, test/test_pytree.py::TestGenericPytree::test_pytree_serialize_bad_input_cxx, test/test_pytree.py::TestGenericPytree::test_pytree_serialize_bad_input_py, test/test_pytree.py::TestGenericPytree::test_register_pytree_node_cxx, test/test_pytree.py::TestGenericPytree::test_register_pytree_node_py, test/test_pytree.py::TestGenericPytree::test_tree_all_any_cxx, test/test_pytree.py::TestGenericPytree::test_tree_all_any_py, test/test_pytree.py::TestGenericPytree::test_tree_map_cxx, test/test_pytree.py::TestGenericPytree::test_tree_map_multi_inputs_cxx, test/test_pytree.py::TestGenericPytree::test_tree_map_multi_inputs_py, test/test_pytree.py::TestGenericPytree::test_tree_map_only_cxx, test/test_pytree.py::TestGenericPytree::test_tree_map_only_predicate_fn_cxx, test/test_pytree.py::TestGenericPytree::test_tree_map_only_predicate_fn_py, test/test_pytree.py::TestGenericPytree::test_tree_map_only_py, test/test_pytree.py::TestGenericPytree::test_tree_map_py, test/test_pytree.py::TestPythonPytree::test_constant, test/test_pytree.py::TestPythonPytree::test_constant_default_eq_error, test/test_pytree.py::TestPythonPytree::test_constant_default_hash_error, test/test_pytree.py::TestPythonPytree::test_dataclass, test/test_pytree.py::TestPythonPytree::test_deprecated_register_pytree_node, test/test_pytree.py::TestPythonPytree::test_flatten_flatten_with_key_consistency, test/test_pytree.py::TestPythonPytree::test_import_pytree_doesnt_import_optree, test/test_pytree.py::TestPythonPytree::test_key_access, test/test_pytree.py::TestPythonPytree::test_key_str, test/test_pytree.py::TestPythonPytree::test_pytree_context_serialize_bad, test/test_pytree.py::TestPythonPytree::test_pytree_custom_type_serialize, test/test_pytree.py::TestPythonPytree::test_pytree_custom_type_serialize_bad, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_bad_protocol, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_defaultdict_enum, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_enum, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_namedtuple, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_namedtuple_bad, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_register_bad, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec0, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec1, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec2, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec3, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec4, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec5, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec6, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec7, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec8, test/test_pytree.py::TestPythonPytree::test_pytree_serialize_spec9, test/test_pytree.py::TestPythonPytree::test_register_dataclass_class, test/test_pytree.py::TestPythonPytree::test_saved_serialized, test/test_pytree.py::TestPythonPytree::test_tree_flatten_with_path_is_leaf, test/test_pytree.py::TestPythonPytree::test_tree_flatten_with_path_roundtrip, test/test_pytree.py::TestPythonPytree::test_tree_leaves_with_path, test/test_pytree.py::TestPythonPytree::test_tree_map_with_path, test/test_pytree.py::TestPythonPytree::test_tree_map_with_path_multiple_trees, test/test_pytree.py::TestPythonPytree::test_treespec_equality, test/test_pytree.py::TestPythonPytree::test_treespec_repr, test/test_pytree.py::TestCxxPytree::test_pytree_custom_type_serialize, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_namedtuple, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec0, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec1, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec2, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec3, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec4, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec5, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec6, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec7, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec8, test/test_pytree.py::TestCxxPytree::test_pytree_serialize_spec9, test/test_pytree.py::TestCxxPytree::test_treespec_equality, test/test_pytree.py::TestCxxPytree::test_treespec_repr 2025-07-17T10:33:46.9713028Z 2025-07-17T10:33:46.9713159Z Running test_quantization 1/3 ... [2025-07-17 10:33:46.967923] 2025-07-17T10:33:46.9713432Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:33:46.9714210Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_quantization.py', '-m', 'not serial', '--shard-id=1', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:33:46.968247] 2025-07-17T10:41:17.2269307Z 2025-07-17T10:41:17.2270048Z test_quantization 1/3 was successful, full logs can be found in artifacts with path test/test-reports/test_quantization_1.3_6b6ef7defeeb5eb5_.log 2025-07-17T10:41:17.2352944Z Running 441 items in this shard: test/test_quantization.py::TestQuantizedOps::test_adaptive_avg_pool3d_ndhwc, test/test_quantization.py::TestQuantizedOps::test_cat, test/test_quantization.py::TestQuantizedOps::test_cat_nhwc, test/test_quantization.py::TestQuantizedOps::test_custom_module_multi_head_attention, test/test_quantization.py::TestQuantizedOps::test_group_norm, test/test_quantization.py::TestQuantizedOps::test_hardswish, test/test_quantization.py::TestQuantizedOps::test_hardtanh, test/test_quantization.py::TestQuantizedOps::test_instance_norm, test/test_quantization.py::TestQuantizedOps::test_int8_add_onednn, test/test_quantization.py::TestQuantizedOps::test_int8_batch_norm_onednn, test/test_quantization.py::TestQuantizedOps::test_int8_mul_onednn, test/test_quantization.py::TestQuantizedOps::test_max_pool2d_nhwc, test/test_quantization.py::TestQuantizedOps::test_mul_scalar_relu, test/test_quantization.py::TestQuantizedOps::test_qadd_relu_cudnn_nhwc, test/test_quantization.py::TestQuantizedOps::test_qadd_relu_same_qparams, test/test_quantization.py::TestQuantizedOps::test_qcelu, test/test_quantization.py::TestQuantizedOps::test_qelu, test/test_quantization.py::TestQuantizedOps::test_qhardsigmoid, test/test_quantization.py::TestQuantizedOps::test_qlayer_norm, test/test_quantization.py::TestQuantizedOps::test_qmul_broadcast, test/test_quantization.py::TestQuantizedOps::test_qmul_relu_different_qparams, test/test_quantization.py::TestQuantizedOps::test_qprelu, test/test_quantization.py::TestQuantizedOps::test_qrelu6, test/test_quantization.py::TestQuantizedOps::test_qtopk, test/test_quantization.py::TestQuantizedOps::test_quantized_equal, test/test_quantization.py::TestQuantizedOps::test_quantized_mean_qnnpack, test/test_quantization.py::TestQuantizedOps::test_sigmoid_dequantize_rounding_error, test/test_quantization.py::TestQuantizedOps::test_sigmoid_non_observed, test/test_quantization.py::TestQuantizedOps::test_std, test/test_quantization.py::TestQNNPackOps::test_qnnpack_add, test/test_quantization.py::TestQNNPackOps::test_qnnpack_mul, test/test_quantization.py::TestQNNPackOps::test_qnnpack_sigmoid, test/test_quantization.py::TestQNNPackOps::test_qnnpack_sigmoid_sweep, test/test_quantization.py::TestQNNPackOps::test_qnnpack_tanh, test/test_quantization.py::TestQuantizedLinear::test_qlinear_add_relu_fp8, test/test_quantization.py::TestQuantizedLinear::test_qlinear_pt2e, test/test_quantization.py::TestQuantizedLinear::test_qlinear_qnnpack_free_memory_and_unpack, test/test_quantization.py::TestQuantizedLinear::test_qlinear_relu, test/test_quantization.py::TestQuantizedLinear::test_qlinear_relu_fp8, test/test_quantization.py::TestQuantizedLinear::test_qlinear_sum_relu_fp8, test/test_quantization.py::TestQuantizedLinear::test_qlinear_sum_relu_pt2e, test/test_quantization.py::TestQuantizedConv::test_conv_transpose_reorder_issue_onednn, test/test_quantization.py::TestQuantizedConv::test_qconv1d_cudnn, test/test_quantization.py::TestQuantizedConv::test_qconv1d_pt2e, test/test_quantization.py::TestQuantizedConv::test_qconv1d_relu_pt2e, test/test_quantization.py::TestQuantizedConv::test_qconv1d_unpack, test/test_quantization.py::TestQuantizedConv::test_qconv2d, test/test_quantization.py::TestQuantizedConv::test_qconv2d_add_relu, test/test_quantization.py::TestQuantizedConv::test_qconv2d_hardswish_fp8, test/test_quantization.py::TestQuantizedConv::test_qconv2d_pt2e, test/test_quantization.py::TestQuantizedConv::test_qconv2d_sum_relu_pt2e, test/test_quantization.py::TestQuantizedConv::test_qconv2d_swish_fp8, test/test_quantization.py::TestQuantizedConv::test_qconv2d_swish_pt2e, test/test_quantization.py::TestQuantizedConv::test_qconv3d, test/test_quantization.py::TestDynamicQuantizedOps::test_dynamic_conv3d, test/test_quantization.py::TestDynamicQuantizedOps::test_dynamic_convtranspose1d, test/test_quantization.py::TestDynamicQuantizedOps::test_linear_prepack_fp16_numerics, test/test_quantization.py::TestDynamicQuantizedOps::test_qlstmGRU, test/test_quantization.py::TestComparatorOps::test_compare_tensor_tensor, test/test_quantization.py::TestPadding::test_reflection_pad1d, test/test_quantization.py::TestPadding::test_reflection_pad2d, test/test_quantization.py::TestQuantizedFunctionalOps::test_conv1d_api, test/test_quantization.py::TestQuantizedFunctionalOps::test_conv3d_api, test/test_quantization.py::TestFakeQuantizeOps::test_backward_per_channel_cachemask_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_fake_quant_control, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_channel, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_channel_cachemask_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_channel_half_precision_numerics, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_tensor, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_forward_per_tensor_cpu, test/test_quantization.py::TestFusedObsFakeQuant::test_fused_obs_fake_quant_backward_op, test/test_quantization.py::TestFusedObsFakeQuant::test_fused_obs_fake_quant_moving_avg_per_channel, test/test_quantization.py::TestQuantizedTensor::test_choose_qparams_optimized, test/test_quantization.py::TestQuantizedTensor::test_decomposed_dequantize_per_tensor, test/test_quantization.py::TestQuantizedTensor::test_decomposed_quantize_per_channel_bfloat16_input, test/test_quantization.py::TestQuantizedTensor::test_decomposed_quantize_per_token, test/test_quantization.py::TestQuantizedTensor::test_dequantize_fp16_cpu, test/test_quantization.py::TestQuantizedTensor::test_jit_serialization, test/test_quantization.py::TestQuantizedTensor::test_per_channel_qtensor_to_memory_format, test/test_quantization.py::TestQuantizedTensor::test_per_tensor_qtensor_to_memory_format, test/test_quantization.py::TestQuantizedTensor::test_qtensor_copy, test/test_quantization.py::TestQuantizedTensor::test_qtensor_creation, test/test_quantization.py::TestQuantizedTensor::test_qtensor_cuda, test/test_quantization.py::TestQuantizedTensor::test_qtensor_dtypes, test/test_quantization.py::TestQuantizedTensor::test_qtensor_fill_per_tensor, test/test_quantization.py::TestQuantizedTensor::test_qtensor_index_put_cpu, test/test_quantization.py::TestQuantizedTensor::test_qtensor_index_put_cuda, test/test_quantization.py::TestQuantizedTensor::test_qtensor_masked_fill_cpu, test/test_quantization.py::TestQuantizedTensor::test_qtensor_masked_fill_cuda, test/test_quantization.py::TestQuantizedTensor::test_qtensor_per_channel_load_save, test/test_quantization.py::TestQuantizedTensor::test_qtensor_permute, test/test_quantization.py::TestQuantizedTensor::test_qtensor_quant_dequant, test/test_quantization.py::TestQuantizedTensor::test_qtensor_sub_byte_aligned_cols, test/test_quantization.py::TestQuantizedTensor::test_qtensor_unsqueeze, test/test_quantization.py::TestQuantizedTensor::test_quantize_per_channel_sub_byte, test/test_quantization.py::TestFakeQuantize::test_fq_module_per_channel, test/test_quantization.py::TestFakeQuantize::test_fq_serializable_per_channel, test/test_quantization.py::TestObserver::test_histogram_observer_ignore_infinity, test/test_quantization.py::TestObserver::test_observer_scriptable, test/test_quantization.py::TestObserver::test_save_load_state_dict_script, test/test_quantization.py::TestStaticQuantizedModule::test_batch_norm2d_serialization, test/test_quantization.py::TestStaticQuantizedModule::test_batch_norm3d, test/test_quantization.py::TestStaticQuantizedModule::test_batch_norm3d_serialization, test/test_quantization.py::TestStaticQuantizedModule::test_conv3d_relu_api, test/test_quantization.py::TestStaticQuantizedModule::test_dropout, test/test_quantization.py::TestStaticQuantizedModule::test_group_norm, test/test_quantization.py::TestStaticQuantizedModule::test_sigmoid, test/test_quantization.py::TestDynamicQuantizedModule::test_dynamic_conv2d, test/test_quantization.py::TestDynamicQuantizedModule::test_dynamic_convtranspose2d, test/test_quantization.py::TestDynamicQuantizedModule::test_lstm_api, test/test_quantization.py::TestReferenceQuantizedModule::test_rnn, test/test_quantization.py::TestHistogramObserver::test_histogram_observer_extreme_inputs, test/test_quantization.py::TestHistogramObserver::test_histogram_observer_one_sided, test/test_quantization.py::TestHistogramObserver::test_histogram_observer_same_inputs, test/test_quantization.py::TestHistogramObserver::test_histogram_observer_update_within_range_succeeds, test/test_quantization.py::TestHistogramObserver::test_observer_scriptable, test/test_quantization.py::TestDistributed::test_qat_convbn_fused_syncbn_replacement, test/test_quantization.py::TestFusedObsFakeQuantModule::test_compare_fused_obs_fq_oss_module, test/test_quantization.py::TestFusedObsFakeQuantModule::test_embedding_qat_config, test/test_quantization.py::TestFusedObsFakeQuantModule::test_fused_mod_per_channel, test/test_quantization.py::TestBackendConfig::test_backend_config_from_dict, test/test_quantization.py::TestBackendConfig::test_backend_config_to_dict, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_extra_inputs_getter, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_fuser_method, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_observation_type, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_qat_module, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_root_module, test/test_quantization.py::TestBackendConfig::test_dtype_config_to_dict, test/test_quantization.py::TestUtils::test_get_fqn_to_example_inputs_default_kwargs, test/test_quantization.py::TestUtils::test_uint4_int4_dtype, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_activations, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_activations_in_non_leaf_module_list, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_convtranspose_per_channel_fails_early, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_convtranspose_per_channel_qconfig_none, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_quantized_embedding, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_quantized_embedding_bag, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_save_load_state_dict, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_embedding_ops_dynamic, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_single_layer, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_type_match_rule, test/test_quantization.py::TestQuantizeEagerOps::test_conv_transpose_3d, test/test_quantization.py::TestQuantizeEagerOps::test_functional_module, test/test_quantization.py::TestQuantizeEagerOps::test_linear, test/test_quantization.py::TestQuantizeEagerQAT::test_add_scalar_uses_input_qparams, test/test_quantization.py::TestQuantizeEagerQAT::test_mul_scalar_uses_input_qparams, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_linear_bn_symm_numerics, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_linear_bn_workflow, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_linear_precomputed_fake_quant, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_relu, test/test_quantization.py::TestFuseEager::test_forward_hooks_preserved, test/test_quantization.py::TestFuseEager::test_fuse_module_train, test/test_quantization.py::TestFuseEager::test_fusion_conv_with_bias, test/test_quantization.py::TestFuseEager::test_fusion_convtranspose_bn_eval, test/test_quantization.py::TestFuseEager::test_fusion_linear_bn_eval, test/test_quantization.py::TestModelNumericsEager::test_fake_quant_true_quant_compare, test/test_quantization.py::TestModelNumericsEager::test_float_quant_compare_per_channel, test/test_quantization.py::TestModelNumericsEager::test_float_quant_compare_per_tensor, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_conv_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_linear_dynamic, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_conv_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_lstm_dynamic, test/test_quantization.py::TestNumericSuiteEager::test_compare_weights_conv_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_weights_lstm_dynamic, test/test_quantization.py::TestNumericSuiteEager::test_output_logger, test/test_quantization.py::TestEqualizeEager::test_converged, test/test_quantization.py::TestEqualizeEager::test_cross_layer_equalization, test/test_quantization.py::TestEqualizeEager::test_equalize, test/test_quantization.py::TestFuseFx::test_fuse_conv_bn_relu, test/test_quantization.py::TestFuseFx::test_fuse_convtranspose_bn_eval, test/test_quantization.py::TestQuantizeFx::test__convert_to_reference_decomposed_fx_dynamic_quant, test/test_quantization.py::TestQuantizeFx::test__convert_to_reference_decomposed_fx_per_channel_quant, test/test_quantization.py::TestQuantizeFx::test_assert_on_size_after_quant_layer, test/test_quantization.py::TestQuantizeFx::test_backend_config_quantization_range, test/test_quantization.py::TestQuantizeFx::test_backend_config_scale_min, test/test_quantization.py::TestQuantizeFx::test_channel_shuffle_lowering, test/test_quantization.py::TestQuantizeFx::test_conv_transpose_reference, test/test_quantization.py::TestQuantizeFx::test_conv_transpose_relu_reference, test/test_quantization.py::TestQuantizeFx::test_convert_custom_config_set_preserved_attributes, test/test_quantization.py::TestQuantizeFx::test_convert_qconfig_mapping, test/test_quantization.py::TestQuantizeFx::test_convtranspose_per_channel_fails_early, test/test_quantization.py::TestQuantizeFx::test_custom_module_class, test/test_quantization.py::TestQuantizeFx::test_deepcopy_preserve_attributes, test/test_quantization.py::TestQuantizeFx::test_dequantize, test/test_quantization.py::TestQuantizeFx::test_dict_output, test/test_quantization.py::TestQuantizeFx::test_dynamic_quant_fp16, test/test_quantization.py::TestQuantizeFx::test_dynamic_quant_weight_observer, test/test_quantization.py::TestQuantizeFx::test_dynamic_with_fusion, test/test_quantization.py::TestQuantizeFx::test_dynamic_with_fusion_multiple_uses, test/test_quantization.py::TestQuantizeFx::test_fold_quant_dequant, test/test_quantization.py::TestQuantizeFx::test_fp32_input_fp32_output, test/test_quantization.py::TestQuantizeFx::test_fuse_custom_config_from_dict, test/test_quantization.py::TestQuantizeFx::test_fuse_custom_config_to_dict, test/test_quantization.py::TestQuantizeFx::test_linear_leaky_relu_lowering, test/test_quantization.py::TestQuantizeFx::test_linear_tanh_lowering, test/test_quantization.py::TestQuantizeFx::test_lowering_functional_conv_transpose_with_kwargs, test/test_quantization.py::TestQuantizeFx::test_lowering_functional_linear_with_kwargs, test/test_quantization.py::TestQuantizeFx::test_masked_fill_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_mul_add_fp16_config, test/test_quantization.py::TestQuantizeFx::test_non_traceable_module, test/test_quantization.py::TestQuantizeFx::test_output_lists_and_dicts, test/test_quantization.py::TestQuantizeFx::test_permute_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_from_dict, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_float_to_observed_mapping, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_non_traceable_module_classes, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_standalone_module_name, test/test_quantization.py::TestQuantizeFx::test_prepare_mode, test/test_quantization.py::TestQuantizeFx::test_prepared_model_deepcopy, test/test_quantization.py::TestQuantizeFx::test_preserve_qconfig, test/test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_split_list_args, test/test_quantization.py::TestQuantizeFx::test_qat_prepare_device_affinity, test/test_quantization.py::TestQuantizeFx::test_qconfig_for_call_func, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_module_name_regex, test/test_quantization.py::TestQuantizeFx::test_quantized_input_fp32_output, test/test_quantization.py::TestQuantizeFx::test_ref_conv_module, test/test_quantization.py::TestQuantizeFx::test_register_patterns, test/test_quantization.py::TestQuantizeFx::test_relu_lowering, test/test_quantization.py::TestQuantizeFx::test_reroute_tuple_getitem_patterns, test/test_quantization.py::TestQuantizeFx::test_reshape_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_return_none, test/test_quantization.py::TestQuantizeFx::test_shape_followed_by_quantized_op, test/test_quantization.py::TestQuantizeFx::test_stack_trace_preserved_linear, test/test_quantization.py::TestQuantizeFx::test_standalone_module_quantized_interface, test/test_quantization.py::TestQuantizeFx::test_state_dict, test/test_quantization.py::TestQuantizeFx::test_sub_scalar, test/test_quantization.py::TestQuantizeFx::test_symmetric_qnnpack_qat_qconfig_mapping, test/test_quantization.py::TestQuantizeFx::test_symmetric_qnnpack_qconfig_mapping, test/test_quantization.py::TestQuantizeFxOps::test_cat, test/test_quantization.py::TestQuantizeFxOps::test_conv_module, test/test_quantization.py::TestQuantizeFxOps::test_conv_transpose_1d, test/test_quantization.py::TestQuantizeFxOps::test_copy_node_fp32_input, test/test_quantization.py::TestQuantizeFxOps::test_div, test/test_quantization.py::TestQuantizeFxOps::test_elu, test/test_quantization.py::TestQuantizeFxOps::test_embedding, test/test_quantization.py::TestQuantizeFxOps::test_embedding_bag, test/test_quantization.py::TestQuantizeFxOps::test_fixed_qparams_ops, test/test_quantization.py::TestQuantizeFxOps::test_fixed_qparams_ops_qint8, test/test_quantization.py::TestQuantizeFxOps::test_gelu_reference, test/test_quantization.py::TestQuantizeFxOps::test_hardswish, test/test_quantization.py::TestQuantizeFxOps::test_instance_norm, test/test_quantization.py::TestQuantizeFxOps::test_int8_input_no_unnecessary_fq, test/test_quantization.py::TestQuantizeFxOps::test_leaky_relu, test/test_quantization.py::TestQuantizeFxOps::test_linear_dynamic_fp16, test/test_quantization.py::TestQuantizeFxOps::test_linear_module, test/test_quantization.py::TestQuantizeFxOps::test_linear_static_fp16, test/test_quantization.py::TestQuantizeFxOps::test_mul, test/test_quantization.py::TestQuantizeFxOps::test_norm_weight_bias, test/test_quantization.py::TestQuantizeFxOps::test_pixel_unshuffle_module, test/test_quantization.py::TestQuantizeFxOps::test_qbatch_norm, test/test_quantization.py::TestQuantizeFxOps::test_qmatmul, test/test_quantization.py::TestQuantizeFxOps::test_quantized_mul_qat, test/test_quantization.py::TestQuantizeFxModels::test_qat_embedding_linear, test/test_quantization.py::TestQuantizeFxModels::test_resnet_base, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_internal_pattern_nodes_cannot_have_users_that_are_not_matched, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_pattern_output_pattern_node_can_have_users_that_are_not_matched, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_placeholder_matching, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_replaces_referenced_submodules, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_single_pattern_match, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_traced_as_callable, test/test_quantization.py::TestGraphUtils::test_conv_bn_conv_relu, test/test_quantization.py::TestGraphUtils::test_conv_bn_relu, test/test_quantization.py::TestDuplicateDQPass::test_avgpool_use_different_qconfig, test/test_quantization.py::TestDuplicateDQPass::test_simple_duplicate_dq, test/test_quantization.py::TestMetaDataPorting::test_metadata_porting_for_dq, test/test_quantization.py::TestMetaDataPorting::test_no_metadata_porting, test/test_quantization.py::TestNumericDebugger::test_control_flow, test/test_quantization.py::TestNumericDebugger::test_extract_results_from_loggers, test/test_quantization.py::TestNumericDebugger::test_prepare_for_propagation_comparison, test/test_quantization.py::TestNumericDebugger::test_re_export_preserve_handle, test/test_quantization.py::TestNumericDebugger::test_run_decompositions_map_handle_to_new_nodes, test/test_quantization.py::TestQuantizePT2E::test_allow_exported_model_train_eval, test/test_quantization.py::TestQuantizePT2E::test_allow_exported_model_train_eval_idempotent, test/test_quantization.py::TestQuantizePT2E::test_constant_prop_preserve_metadata, test/test_quantization.py::TestQuantizePT2E::test_derived_qspec_per_channel, test/test_quantization.py::TestQuantizePT2E::test_fixed_qparams_qspec_ptq, test/test_quantization.py::TestQuantizePT2E::test_fold_all_ops_before_quantize, test/test_quantization.py::TestQuantizePT2E::test_max_pool2d_quantizer, test/test_quantization.py::TestQuantizePT2E::test_model_is_exported, test/test_quantization.py::TestQuantizePT2E::test_move_exported_model_dropout, test/test_quantization.py::TestQuantizePT2E::test_move_exported_model_dropout_inplace, test/test_quantization.py::TestQuantizePT2E::test_multi_users_without_output_observer, test/test_quantization.py::TestQuantizePT2E::test_quantization_dtype_bfloat16_int16, test/test_quantization.py::TestQuantizePT2E::test_simple_quantizer, test/test_quantization.py::TestPT2ERepresentation::test_add, test/test_quantization.py::TestPT2ERepresentation::test_add_relu, test/test_quantization.py::TestPT2ERepresentation::test_maxpool2d, test/test_quantization.py::TestPT2ERepresentation::test_qdq, test/test_quantization.py::TestPT2ERepresentation::test_static_linear, test/test_quantization.py::TestXNNPACKQuantizer::test_add_and_inplace_add, test/test_quantization.py::TestXNNPACKQuantizer::test_conv_linear, test/test_quantization.py::TestXNNPACKQuantizer::test_linear, test/test_quantization.py::TestXNNPACKQuantizer::test_propagate_annotation, test/test_quantization.py::TestXNNPACKQuantizer::test_set_module_name, test/test_quantization.py::TestXNNPACKQuantizer::test_set_module_name_with_underscores, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_adaptive_avg_pool2d_recipe, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_avg_pool2d_recipe, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_cat_recipe_single_input, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_conv2d_serials_binary_unary, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_conv2d_unary, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_dynamic_quant_linear, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_filter_linear_recipe, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_flatten_recipe2, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_linear_binary, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_linear_binary_unary_dynamic, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_linear_dynamic_fp16, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_qat_dynamic_quant_linear, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_set_module_name_and_module_type_with_mixed_configs, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_conv_bn_fusion, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_conv_bn_per_channel_weight_bias, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_conv_bn_relu_fusion, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_conv_bn_relu_fusion_cuda, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_conv_bn_relu_fusion_no_conv_bias, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_conv_transpose_bn, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_inplace_add_relu, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_per_channel_weight_custom_dtype, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_update_shared_qspec, test/test_quantization.py::TestQuantizePT2EQATModels::test_qat_mobilenet_v2, test/test_quantization.py::TestQuantizePT2EQATModels::test_qat_resnet18, test/test_quantization.py::TestFXGraphMatcher::test_matching_failure_node_count, test/test_quantization.py::TestFXGraphMatcher::test_methods, test/test_quantization.py::TestFXGraphMatcher::test_results_order, test/test_quantization.py::TestFXGraphMatcher::test_simple_mod, test/test_quantization.py::TestFXGraphMatcher::test_simple_mod_multi, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_loggers_cuda, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_mul_inputs_activations, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_mod_qat, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extend_logger_results_with_comparison, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_cuda, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_int8_shadows_fp32_coverage, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_layer_names, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_meth_ptq, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_mod_qat, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_op_io_dtype_coverage, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_op_with_either_fp32_or_int8_input, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_op_with_only_kwargs_skips_shadowing, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_ops_with_same_fp32_and_int8_signature, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_user_module_scriptable, test/test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_linear_mod_fp32_fp32, test/test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_linear_mod_quant_fp32, test/test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_linear_mod_quant_quant, test/test_quantization.py::TestFXNumericSuiteNShadows::test_partial_qconfig_mapping, test/test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_deduplication, test/test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_end_to_end, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_activations_linear, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_weights_linear, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_weights_lstm_dynamic, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_sparsenn_compare_activations, test/test_quantization.py::TestFxModelReportDetector::test_conv_sub_class_considered, test/test_quantization.py::TestFxModelReportDetector::test_simple_conv, test/test_quantization.py::TestFxModelReportObserver::test_single_batch_of_ones, test/test_quantization.py::TestFxModelReportClass::test_equalization_mapping_generation, test/test_quantization.py::TestFxModelReportClass::test_generate_report, test/test_quantization.py::TestFxModelReportClass::test_qconfig_mapping_generation, test/test_quantization.py::TestFxDetectInputWeightEqualization::test_input_weight_equalization_report_gen, test/test_quantization.py::TestFxDetectOutliers::test_all_outlier_report_gen, test/test_quantization.py::TestFxDetectOutliers::test_no_outlier_report_gen, test/test_quantization.py::TestFxModelReportVisualizer::test_generate_tables_match_with_report, test/test_quantization.py::TestFxModelReportVisualizer::test_generate_tables_single_feat_match, test/test_quantization.py::TestFxModelReportVisualizer::test_get_modules_and_features, test/test_quantization.py::TestEqualizeFx::test_input_weight_equalization_equalization_scales, test/test_quantization.py::TestEqualizeFx::test_selective_equalization, test/test_quantization.py::TestSerialization::test_conv2d, test/test_quantization.py::TestSerialization::test_conv2d_graph_v2, test/test_quantization.py::TestSerialization::test_conv2d_nobias, test/test_quantization.py::TestSerialization::test_conv2d_relu, test/test_quantization.py::TestSerialization::test_conv3d, test/test_quantization.py::TestSerialization::test_conv3d_relu, test/test_quantization.py::TestSerialization::test_linear, test/test_quantization.py::TestSerialization::test_linear_dynamic, test/test_quantization.py::TestSerialization::test_linear_relu, test/test_quantization.py::TestSerialization::test_lstm, test/test_quantization.py::TestSerialization::test_per_channel_observer, test/test_quantization.py::TestSerialization::test_per_tensor_observer, test/test_quantization.py::TestQuantizeJit::test_conv_transpose, test/test_quantization.py::TestQuantizeJit::test_linear_dynamic_fp16, test/test_quantization.py::TestQuantizeJit::test_nested, test/test_quantization.py::TestQuantizeJit::test_skip_quant, test/test_quantization.py::TestQuantizeJitPasses::test_finalize_debug, test/test_quantization.py::TestQuantizeJitPasses::test_finalize_for_linear, test/test_quantization.py::TestQuantizeJitPasses::test_foldbn_in_submodule, test/test_quantization.py::TestQuantizeJitPasses::test_foldbn_no_fusion, test/test_quantization.py::TestQuantizeJitPasses::test_foldbn_shared_classtype, test/test_quantization.py::TestQuantizeJitPasses::test_foldbn_trivial, test/test_quantization.py::TestQuantizeJitPasses::test_foldbn_trivial_nobias, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_propagate_observed, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_shared_class_type, test/test_quantization.py::TestQuantizeJitPasses::test_insert_quant_dequant, test/test_quantization.py::TestQuantizeJitPasses::test_interface_with_fork, test/test_quantization.py::TestQuantizeJitPasses::test_module_list, test/test_quantization.py::TestQuantizeJitPasses::test_replicate_dequantize_in_block, test/test_quantization.py::TestQuantizeJitPasses::test_replicate_quantize_for_if, test/test_quantization.py::TestQuantizeJitPasses::test_skip_dequant_constant_prop, test/test_quantization.py::TestQuantizeJitPasses::test_swap_functional_linear, test/test_quantization.py::TestQuantizeJitOps::test_dequantize_tuple, test/test_quantization.py::TestQuantizeJitOps::test_instance_norm, test/test_quantization.py::TestQuantizeJitOps::test_qbatch_norm_relu_BNFuncInplaceRelu, test/test_quantization.py::TestQuantizeJitOps::test_quantized_add_relu_alpha, test/test_quantization.py::TestQuantizeJitOps::test_quantized_add_scalar, test/test_quantization.py::TestQuantizeJitOps::test_quantized_cat, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_multi_op, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_weight_observer, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_prepare_dynamic, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_quantize_dynamic_fp16, test/test_quantization.py::TestQuantizeDynamicJitOps::test_embedding_bag, test/test_quantization.py::TestDeprecatedJitQuantized::test_erase_class_tensor_shapes, test/test_quantization.py::TestDeprecatedJitQuantized::test_quantization_modules, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_fuse_modules, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_fuser_method_mappings, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_stubs, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_embedding_ops, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_normalization, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_import_nn_intrinsic, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_import_nn_intrinsic_quantized, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_qat_linear_fused, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_quantized_conv_relu, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_quantized_linear_relu, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_nn_intrinsic_fused, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_no_import_nn_intrinsic_quantized_dynamic, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_equalize, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_fuse, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_match_utils, test/test_quantization.py::TestBitsCUDA::test_cat_cuda, test/test_quantization.py::TestFloat8DtypeCUDA::test_cast_round_trip_extremes_cuda_float8_e5m2, test/test_quantization.py::TestFloat8DtypeCUDA::test_cast_round_trip_rte_cuda_float8_e4m3fn, test/test_quantization.py::TestFloat8DtypeCUDA::test_cast_round_trip_rte_cuda_float8_e5m2, test/test_quantization.py::TestFloat8DtypeCUDA::test_cast_round_trip_subnormals_cuda_float8_e8m0fnu, test/test_quantization.py::TestFloat8DtypeCUDA::test_cat_cuda_float8_e5m2, test/test_quantization.py::TestFloat8DtypeCUDA::test_empty_cuda_float8_e4m3fn, test/test_quantization.py::TestFloat8DtypeCUDA::test_finfo_cuda_float8_e4m3fn, test/test_quantization.py::TestFloat8DtypeCUDA::test_finfo_cuda_float8_e4m3fnuz, test/test_quantization.py::TestFloat8DtypeCUDA::test_finfo_cuda_float8_e5m2, test/test_quantization.py::TestFloat8DtypeCUDA::test_finfo_cuda_float8_e5m2fnuz, test/test_quantization.py::TestFloat8DtypeCUDA::test_finfo_cuda_float8_e8m0fnu, test/test_quantization.py::TestFloat8DtypeCUDA::test_save_load_cuda_float8_e8m0fnu, test/test_quantization.py::TestFloat8DtypeCUDA::test_special_numbers_cuda_float8_e8m0fnu, test/test_quantization.py::TestFloat8DtypeCUDA::test_to_string_cuda_float8_e4m3fn, test/test_quantization.py::TestFloat8DtypeCUDA::test_to_string_cuda_float8_e5m2, test/test_quantization.py::TestFloat8DtypeCUDA::test_type_promotion_fails_cuda_float8_e4m3fn, test/test_quantization.py::TestFloat8DtypeCUDA::test_type_promotion_fails_cuda_float8_e5m2, test/test_quantization.py::TestFloat8DtypeCUDA::test_type_promotion_fails_cuda_float8_e8m0fnu 2025-07-17T10:41:17.2436408Z 2025-07-17T10:41:17.2436579Z GITHUB_RUN_ID, GITHUB_RUN_ATTEMPT, or ARTIFACTS_FILE_SUFFIX not set, not uploading 2025-07-17T10:41:17.2436921Z Running test_quantization 3/3 ... [2025-07-17 10:41:17.227239] 2025-07-17T10:41:17.2437191Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-07-17T10:41:17.2437410Z Uploading artifacts took 0.00 seconds 2025-07-17T10:41:17.2438202Z Executing ['/opt/conda/envs/py_3.12/bin/python', '-bb', 'test_quantization.py', '-m', 'not serial', '--shard-id=3', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '-x', '--reruns=2', '--import-slow-tests', '--import-disabled-tests'] ... [2025-07-17 10:41:17.227566] 2025-07-17T10:41:34.0561547Z 2025-07-17T10:41:34.0562355Z inductor/test_torchinductor_opinfo 11/13 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_11.13_7cd31d5b7938852e_.log 2025-07-17T10:41:34.0639676Z Running 277 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rmod___cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___ror___cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive___rpow___cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__chunk_cat_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__chunk_cat_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__native_batch_norm_legit_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__unsafe_masked_index_put_accumulate_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive__unsafe_masked_index_put_accumulate_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_abs_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_abs_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_acos_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_add_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_addmv_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_all_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_amax_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_aminmax_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_aminmax_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_any_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_arange_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_argmax_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_argmin_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_argsort_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_as_strided_copy_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_asin_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_asin_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atan_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atanh_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_atleast_2d_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bfloat16_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bitwise_and_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bitwise_or_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bitwise_xor_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_block_diag_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bmm_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_broadcast_to_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bucketize_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_bucketize_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_byte_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cholesky_solve_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_clone_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_conj_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_constant_pad_nd_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_constant_pad_nd_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_copysign_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_corrcoef_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cov_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cummax_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cumsum_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_cumsum_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_deg2rad_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diag_embed_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_diag_embed_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_dist_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_div_floor_rounding_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_dot_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_dsplit_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_dsplit_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_empty_permuted_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_eq_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_eq_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_erfc_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_erfinv_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_exp2_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_exp2_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_expand_copy_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_expm1_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fft_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fft_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fftn_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_fftshift_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_hfft2_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_hfft2_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_hfft_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ihfft2_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ihfft_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_ihfft_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_irfft2_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_irfft_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_irfftn_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_irfftn_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_rfft_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fft_rfftn_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fill_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_flatten_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_fliplr_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_flipud_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_float_power_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_floor_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_full_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_full_like_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_geqrf_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_hypot_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_igamma_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_igammac_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_put_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_index_select_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_inner_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isfinite_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isnan_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isneginf_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_isreal_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_item_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_2inputs_2outputs_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_jiterator_4inputs_with_extra_args_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_kron_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ldexp_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_le_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_lgamma_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_eigh_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_ldl_factor_ex_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_ldl_factor_ex_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_lu_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_matrix_norm_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_norm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_norm_subgradients_at_zero_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_pinv_singular_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_svdvals_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_vander_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_vecdot_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linalg_vector_norm_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_linspace_tensor_overload_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log10_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log_softmax_with_dtype_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_log_softmax_with_dtype_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logical_and_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_logical_xor_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_lt_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_lt_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_lu_unpack_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mH_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_argmin_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_cumsum_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_norm_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_normalize_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_prod_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_scatter_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_masked_select_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_matrix_exp_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_max_binary_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_max_pool2d_with_indices_backward_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_max_reduction_no_dim_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_max_reduction_no_dim_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_meshgrid_list_of_tensors_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_binary_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_binary_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_binary_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_min_reduction_with_dim_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_minimum_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mode_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mode_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mul_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_multinomial_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_multinomial_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_mvlgamma_mvlgamma_p_5_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nan_to_num_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nanquantile_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_narrow_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_native_batch_norm_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ne_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ne_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_neg_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_neg_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_new_full_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_max_pool1d_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_avg_pool2d_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_avg_pool3d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_batch_norm_without_cudnn_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_channel_shuffle_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_conv1d_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_conv3d_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_ctc_loss_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_dropout2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_embedding_bag_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_fractional_max_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_instance_norm_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_linear_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_nearest-exact_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_l1_loss_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_max_unpool1d_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_circular_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_constant_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_replicate_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_pad_replicate_negative_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_relu6_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_scaled_dot_product_attention_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_triplet_margin_loss_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nonzero_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_ones_like_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_1_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_polygamma_polygamma_n_1_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_prod_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rad2deg_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rad2deg_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_rad2deg_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_randn_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_reciprocal_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_reciprocal_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_repeat_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_repeat_interleave_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_repeat_interleave_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_resize_as__cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_resolve_conj_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_round_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scalar_tensor_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scalar_tensor_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_add_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_add_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_reduce_amin_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_reduce_prod_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_scatter_reduce_sum_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_select_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_select_scatter_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_short_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_short_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_signal_windows_bartlett_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sinc_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_slice_scatter_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_softmax_with_dtype_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_bessel_j1_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_bessel_j1_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_chebyshev_polynomial_u_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_chebyshev_polynomial_u_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_hermite_polynomial_h_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_i0e_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_i1_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_i1e_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_laguerre_polynomial_l_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_modified_bessel_k0_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_modified_bessel_k1_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_polygamma_special_polygamma_n_0_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_scaled_modified_bessel_k0_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_scaled_modified_bessel_k0_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_scaled_modified_bessel_k1_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_shifted_chebyshev_polynomial_v_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_spherical_bessel_j0_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_spherical_bessel_j0_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_special_spherical_bessel_j0_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_list_args_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_split_with_sizes_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stack_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_std_mean_unbiased_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_sum_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_svd_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tan_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tanh_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tensor_split_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_tile_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_topk_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_torch_ops_aten__safe_softmax_default_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_torch_ops_aten__safe_softmax_default_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_transpose_copy_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_transpose_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_transpose_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trapezoid_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_trapz_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_triu_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unbind_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unflatten_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unfold_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unfold_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unfold_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unique_consecutive_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unique_cuda_int64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unique_cuda_uint32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unsafe_chunk_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unsqueeze_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_unsqueeze_cuda_int32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_var_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_var_mean_unbiased_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_var_mean_unbiased_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_vdot_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_where_cuda_uint8, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_xlogy_cuda_bool, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_zero__cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_zeros_like_cuda_int32 2025-07-17T10:41:34.0717216Z 2025-07-17T10:51:17.7113760Z 2025-07-17T10:51:17.7115058Z test_quantization 3/3 was successful, full logs can be found in artifacts with path test/test-reports/test_quantization_3.3_d4ab7f60c90dc0df_.log 2025-07-17T10:51:17.7191091Z Running 407 items in this shard: test/test_quantization.py::TestQuantizedOps::test_adaptive_avg_pool, test/test_quantization.py::TestQuantizedOps::test_advanced_indexing, test/test_quantization.py::TestQuantizedOps::test_batch_norm, test/test_quantization.py::TestQuantizedOps::test_batch_norm_relu, test/test_quantization.py::TestQuantizedOps::test_equal, test/test_quantization.py::TestQuantizedOps::test_interpolate, test/test_quantization.py::TestQuantizedOps::test_interpolate3d, test/test_quantization.py::TestQuantizedOps::test_linear_bias_unpack, test/test_quantization.py::TestQuantizedOps::test_max_pool2d_cudnn, test/test_quantization.py::TestQuantizedOps::test_max_pool3d, test/test_quantization.py::TestQuantizedOps::test_max_pool3d_nhwc, test/test_quantization.py::TestQuantizedOps::test_qadd_broadcast, test/test_quantization.py::TestQuantizedOps::test_qadd_relu_different_qparams, test/test_quantization.py::TestQuantizedOps::test_qclamp, test/test_quantization.py::TestQuantizedOps::test_qgelu, test/test_quantization.py::TestQuantizedOps::test_qsoftmax, test/test_quantization.py::TestQuantizedOps::test_qtanh, test/test_quantization.py::TestQNNPackOps::test_adaptive_avg_pool2d, test/test_quantization.py::TestQNNPackOps::test_hardtanh, test/test_quantization.py::TestQNNPackOps::test_qnnpack_relu, test/test_quantization.py::TestQuantizedLinear::test_qlinear, test/test_quantization.py::TestQuantizedLinear::test_qlinear_add_fp8, test/test_quantization.py::TestQuantizedLinear::test_qlinear_add_pt2e, test/test_quantization.py::TestQuantizedLinear::test_qlinear_leaky_relu, test/test_quantization.py::TestQuantizedLinear::test_qlinear_sum_pt2e, test/test_quantization.py::TestQuantizedLinear::test_wrapped_quantized_linear_prepacked, test/test_quantization.py::TestQuantizedConv::test_benchmark, test/test_quantization.py::TestQuantizedConv::test_conv_reorder_issue_onednn, test/test_quantization.py::TestQuantizedConv::test_qconv1d, test/test_quantization.py::TestQuantizedConv::test_qconv1d_fp8, test/test_quantization.py::TestQuantizedConv::test_qconv1d_relu_cudnn, test/test_quantization.py::TestQuantizedConv::test_qconv2d_fp8, test/test_quantization.py::TestQuantizedConv::test_qconv2d_relu_cudnn, test/test_quantization.py::TestQuantizedConv::test_qconv2d_relu_fp8, test/test_quantization.py::TestQuantizedConv::test_qconv2d_relu_pt2e, test/test_quantization.py::TestQuantizedConv::test_qconv2d_sum_fp8, test/test_quantization.py::TestQuantizedConv::test_qconv2d_sum_relu_float_output_pt2e, test/test_quantization.py::TestQuantizedConv::test_qconv2d_sum_relu_fp8, test/test_quantization.py::TestQuantizedConv::test_qconv2d_unpack, test/test_quantization.py::TestQuantizedConv::test_qconv_transpose1d, test/test_quantization.py::TestQuantizedConv::test_qconv_transpose3d, test/test_quantization.py::TestDynamicQuantizedOps::test_dynamic_conv1d, test/test_quantization.py::TestDynamicQuantizedOps::test_dynamic_conv2d, test/test_quantization.py::TestDynamicQuantizedOps::test_dynamic_convtranspose2d, test/test_quantization.py::TestDynamicQuantizedOps::test_dynamic_convtranspose3d, test/test_quantization.py::TestDynamicQuantizedOps::test_qlinear_dynamic_fp16, test/test_quantization.py::TestDynamicQuantizedOps::test_qrnncell, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_2d_indices, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_2bit, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_4bit, test/test_quantization.py::TestQuantizedEmbeddingOps::test_embedding_bag_byte, test/test_quantization.py::TestFakeQuantizeOps::test_backward_per_channel_cachemask_cpu, test/test_quantization.py::TestFakeQuantizeOps::test_backward_per_tensor_cachemask_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_fake_quantize_per_channel_affine_scale_dtypes, test/test_quantization.py::TestFakeQuantizeOps::test_forward_backward_per_tensor_with_amp, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_tensor_cachemask_cpu, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_tensor_cachemask_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_forward_per_tensor_half_precision_numerics, test/test_quantization.py::TestFakeQuantizeOps::test_fq_module_per_tensor, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_backward_per_channel_cpu, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_backward_per_channel_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_backward_per_tensor_cuda, test/test_quantization.py::TestFakeQuantizeOps::test_learnable_forward_per_tensor_cuda, test/test_quantization.py::TestQuantizedTensor::test_bfp16_quantize, test/test_quantization.py::TestQuantizedTensor::test_choose_qparams, test/test_quantization.py::TestQuantizedTensor::test_clone, test/test_quantization.py::TestQuantizedTensor::test_compare_per_tensor_device_numerics, test/test_quantization.py::TestQuantizedTensor::test_cuda_quantization_does_not_pin_memory, test/test_quantization.py::TestQuantizedTensor::test_decomposed_dynamic_quant_pattern, test/test_quantization.py::TestQuantizedTensor::test_decomposed_quantize_per_channel_group, test/test_quantization.py::TestQuantizedTensor::test_decomposed_quantize_per_tensor_bfloat16_input, test/test_quantization.py::TestQuantizedTensor::test_qtensor_channel_float_assignment, test/test_quantization.py::TestQuantizedTensor::test_qtensor_fill_per_channel, test/test_quantization.py::TestQuantizedTensor::test_qtensor_fill_per_channel_nhwc, test/test_quantization.py::TestQuantizedTensor::test_qtensor_index_select_cuda, test/test_quantization.py::TestQuantizedTensor::test_qtensor_load_save, test/test_quantization.py::TestQuantizedTensor::test_qtensor_per_channel_permute, test/test_quantization.py::TestQuantizedTensor::test_qtensor_resize, test/test_quantization.py::TestQuantizedTensor::test_quant_pin_memory, test/test_quantization.py::TestQuantizedTensor::test_repeat, test/test_quantization.py::TestQuantizedTensor::test_torch_qtensor_deepcopy, test/test_quantization.py::TestFakeQuantize::test_quant_min_max_override, test/test_quantization.py::TestObserver::test_histogram_observer_consistent_buffer_shape, test/test_quantization.py::TestObserver::test_histogram_observer_handle_OOM_due_to_close_min_max_value, test/test_quantization.py::TestObserver::test_histogram_observer_save_load_state_dict, test/test_quantization.py::TestObserver::test_per_channel_observers_load_state_dict, test/test_quantization.py::TestObserver::test_state_dict_respects_device_affinity, test/test_quantization.py::TestStaticQuantizedModule::test_batch_norm2d, test/test_quantization.py::TestStaticQuantizedModule::test_conv1d_relu_api, test/test_quantization.py::TestStaticQuantizedModule::test_conv2d_add, test/test_quantization.py::TestStaticQuantizedModule::test_conv3d_api, test/test_quantization.py::TestStaticQuantizedModule::test_embedding_api, test/test_quantization.py::TestStaticQuantizedModule::test_leaky_relu, test/test_quantization.py::TestStaticQuantizedModule::test_linear, test/test_quantization.py::TestStaticQuantizedModule::test_linear_tanh, test/test_quantization.py::TestStaticQuantizedModule::test_prelu, test/test_quantization.py::TestStaticQuantizedModule::test_relu, test/test_quantization.py::TestDynamicQuantizedModule::test_dynamic_conv1d, test/test_quantization.py::TestDynamicQuantizedModule::test_dynamic_conv3d, test/test_quantization.py::TestDynamicQuantizedModule::test_dynamic_convtranspose3d, test/test_quantization.py::TestReferenceQuantizedModule::test_linear_decomposed_weight_custom_qmin_qmax, test/test_quantization.py::TestReferenceQuantizedModule::test_rnn_cell, test/test_quantization.py::TestRecordHistogramObserver::test_observer_scriptable, test/test_quantization.py::TestHistogramObserver::test_histogram_observer, test/test_quantization.py::TestHistogramObserver::test_histogram_observer_against_reference, test/test_quantization.py::TestHistogramObserver::test_histogram_observer_correct_numel, test/test_quantization.py::TestDistributed::test_fake_quant_preserves_buffers, test/test_quantization.py::TestDistributed::test_observers_preserve_buffers, test/test_quantization.py::TestBackendConfig::test_backend_config_set_backend_pattern_config, test/test_quantization.py::TestBackendConfig::test_backend_config_set_name, test/test_quantization.py::TestBackendConfig::test_backend_op_config_from_dict, test/test_quantization.py::TestBackendConfig::test_backend_op_config_set_root_node_getter, test/test_quantization.py::TestBackendConfig::test_backend_op_config_to_dict, test/test_quantization.py::TestUtils::test_get_fqn_to_example_inputs_complex_args, test/test_quantization.py::TestUtils::test_get_fqn_to_example_inputs_simple, test/test_quantization.py::TestUtils::test_quantize_weight_clamping_per_channel, test/test_quantization.py::TestUtils::test_quantize_weight_clamping_per_tensor, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_dequant_stub, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_forward_hooks_preserved, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_manual, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_mha_batch_first_attr_is_copied_in_prepare, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_nested2, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_nested3, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_skip_quant, test/test_quantization.py::TestQuantizeEagerPTQStatic::test_two_layers, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_linear_relu_fusion, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_nested2, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_per_channel_linear_quantize, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_quantized_rnn_cell, test/test_quantization.py::TestQuantizeEagerPTQDynamic::test_two_layers, test/test_quantization.py::TestQuantizeEagerOps::test_conv_1d, test/test_quantization.py::TestQuantizeEagerOps::test_conv_3d, test/test_quantization.py::TestQuantizeEagerOps::test_leaky_relu, test/test_quantization.py::TestQuantizeEagerOps::test_relu, test/test_quantization.py::TestQuantizeEagerQAT::test_dropout, test/test_quantization.py::TestQuantizeEagerQAT::test_dynamic_qat_linear, test/test_quantization.py::TestQuantizeEagerQAT::test_embedding_qat_qconfig_equal, test/test_quantization.py::TestQuantizeEagerQAT::test_eval_only_fake_quant, test/test_quantization.py::TestQuantizeEagerQAT::test_manual, test/test_quantization.py::TestQuantizeEagerQAT::test_train_save_load_eval, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_conv_bn_relu, test/test_quantization.py::TestQuantizeEagerQATNumerics::test_fixed_qparam_ops, test/test_quantization.py::TestFuseEager::test_fuse_function_customization, test/test_quantization.py::TestFuseEager::test_fuse_module_eval, test/test_quantization.py::TestFuseEager::test_fuse_modules_with_nested_hooks, test/test_quantization.py::TestFuseEager::test_fusion_sequential_model_eval, test/test_quantization.py::TestFuseEager::test_fusion_sequential_model_train, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_outputs_linear_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_functional_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_model_stub_linear_static, test/test_quantization.py::TestNumericSuiteEager::test_compare_weights_linear_dynamic, test/test_quantization.py::TestNumericSuiteEager::test_compare_weights_linear_static, test/test_quantization.py::TestNumericSuiteEager::test_mobilenet_v2, test/test_quantization.py::TestEqualizeEager::test_equalize_fused_convrelu, test/test_quantization.py::TestEqualizeEager::test_equalize_fused_linearrelu, test/test_quantization.py::TestFuseFx::test_fuse_addtional_fuser_method, test/test_quantization.py::TestFuseFx::test_fuse_conv_bn_add_relu_lowering, test/test_quantization.py::TestFuseFx::test_fuse_linear_bn_eval, test/test_quantization.py::TestFuseFx::test_fuse_linear_tanh_for_onednn_backend, test/test_quantization.py::TestFuseFx::test_fuse_module_relu, test/test_quantization.py::TestFuseFx::test_fusion_pattern_with_matchallnode, test/test_quantization.py::TestFuseFx::test_fusion_pattern_with_multiple_inputs, test/test_quantization.py::TestFuseFx::test_problematic_fuse_example, test/test_quantization.py::TestFuseFx::test_qconfig_fused_module, test/test_quantization.py::TestQuantizeFx::test__convert_to_reference_decomposed_fx, test/test_quantization.py::TestQuantizeFx::test_conv_linear_not_reference, test/test_quantization.py::TestQuantizeFx::test_conv_lowering, test/test_quantization.py::TestQuantizeFx::test_convert_custom_config_from_dict, test/test_quantization.py::TestQuantizeFx::test_convert_custom_config_set_observed_to_quantized_mapping, test/test_quantization.py::TestQuantizeFx::test_convert_custom_config_to_dict, test/test_quantization.py::TestQuantizeFx::test_custom_module_class_input_has_duplicate_nodes, test/test_quantization.py::TestQuantizeFx::test_default_qconfig_mapping_override_global, test/test_quantization.py::TestQuantizeFx::test_dynamic_linear_input_multiple_use, test/test_quantization.py::TestQuantizeFx::test_get_executorch_backend_config, test/test_quantization.py::TestQuantizeFx::test_linear_qint8_activation, test/test_quantization.py::TestQuantizeFx::test_linear_shape_view, test/test_quantization.py::TestQuantizeFx::test_lowering_functional_conv_with_kwargs, test/test_quantization.py::TestQuantizeFx::test_match_pattern_with_multiple_args, test/test_quantization.py::TestQuantizeFx::test_not_used, test/test_quantization.py::TestQuantizeFx::test_packed_weight_fused_op, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_input_quantized_indexes, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_non_traceable_module_names, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_set_output_quantized_indexes, test/test_quantization.py::TestQuantizeFx::test_prepare_custom_config_to_dict, test/test_quantization.py::TestQuantizeFx::test_preserve_attributes, test/test_quantization.py::TestQuantizeFx::test_preserve_tuple, test/test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_dict_tuple_args, test/test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_list_args, test/test_quantization.py::TestQuantizeFx::test_propagate_dtypes_for_known_nodes_tuple_args, test/test_quantization.py::TestQuantizeFx::test_qat_and_script, test/test_quantization.py::TestQuantizeFx::test_qat_skip_untraced, test/test_quantization.py::TestQuantizeFx::test_qconfig_dict_setup, test/test_quantization.py::TestQuantizeFx::test_qconfig_function, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_repr, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_global, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_set_module_name_object_type_order, test/test_quantization.py::TestQuantizeFx::test_qconfig_mapping_to_dict, test/test_quantization.py::TestQuantizeFx::test_qconfig_none, test/test_quantization.py::TestQuantizeFx::test_qconfig_precedence, test/test_quantization.py::TestQuantizeFx::test_qparams_buffers, test/test_quantization.py::TestQuantizeFx::test_quant_output_always_observed, test/test_quantization.py::TestQuantizeFx::test_quantized_model_type, test/test_quantization.py::TestQuantizeFx::test_ref_linear_module, test/test_quantization.py::TestQuantizeFx::test_save_observer_state_dict, test/test_quantization.py::TestQuantizeFx::test_size_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_standalone_module_float_interface, test/test_quantization.py::TestQuantizeFx::test_static_lstm_consume_tuple, test/test_quantization.py::TestQuantizeFx::test_static_lstm_with_custom_fixed_qparams, test/test_quantization.py::TestQuantizeFx::test_torch_unsqueeze_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_trace_quantize_per_tensor, test/test_quantization.py::TestQuantizeFx::test_unsqueeze__nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFx::test_view_nontensor_args_not_observed, test/test_quantization.py::TestQuantizeFxOps::test_add_relu, test/test_quantization.py::TestQuantizeFxOps::test_add_relu_multiple_uses_of_relu, test/test_quantization.py::TestQuantizeFxOps::test_bmm, test/test_quantization.py::TestQuantizeFxOps::test_clamp, test/test_quantization.py::TestQuantizeFxOps::test_conv_transpose_2d, test/test_quantization.py::TestQuantizeFxOps::test_fixed_qparams_ops_wrong_qconfig, test/test_quantization.py::TestQuantizeFxOps::test_functional_conv, test/test_quantization.py::TestQuantizeFxOps::test_functional_linear, test/test_quantization.py::TestQuantizeFxOps::test_getitem, test/test_quantization.py::TestQuantizeFxOps::test_narrow, test/test_quantization.py::TestQuantizeFxOps::test_pixel_unshuffle, test/test_quantization.py::TestQuantizeFxOps::test_ref_pattern_multi_use, test/test_quantization.py::TestQuantizeFxOps::test_rnn, test/test_quantization.py::TestQuantizeFxOps::test_softmax_reference, test/test_quantization.py::TestQuantizeFxModels::test_model_dropout, test/test_quantization.py::TestQuantizeFxModels::test_switch_device_prepare_convert, test/test_quantization.py::TestQuantizeFxModels::test_torchvision, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_graph_argument_order, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_multiple_pattern_match, test/test_quantization.py::TestSubgraphRewriter::test_subgraph_rewriter_preserves_logic, test/test_quantization.py::TestGraphUtils::test_customized_equivalet_types_dict, test/test_quantization.py::TestDuplicateDQPass::test_no_add_quant_duplicate_dq, test/test_quantization.py::TestMetaDataPorting::test_metadata_porting_for_dq_no_static_q, test/test_quantization.py::TestMetaDataPorting::test_metadata_porting_for_two_dq, test/test_quantization.py::TestMetaDataPorting::test_no_metadata_porting_through_unknown_ops, test/test_quantization.py::TestMetaDataPorting::test_simple_metadata_porting, test/test_quantization.py::TestNumericDebugger::test_added_node_gets_unique_id, test/test_quantization.py::TestNumericDebugger::test_simple, test/test_quantization.py::TestQuantizePT2E::test_allow_implicit_sharing, test/test_quantization.py::TestQuantizePT2E::test_composable_quantizer_throw, test/test_quantization.py::TestQuantizePT2E::test_composable_quantizer_transform_for_annotation, test/test_quantization.py::TestQuantizePT2E::test_conv_padding_bn_relu, test/test_quantization.py::TestQuantizePT2E::test_conv_transpose_bn_relu, test/test_quantization.py::TestQuantizePT2E::test_dont_fold_other_constant, test/test_quantization.py::TestQuantizePT2E::test_embedding_conv_linear_quantization, test/test_quantization.py::TestQuantizePT2E::test_embedding_quantizer, test/test_quantization.py::TestQuantizePT2E::test_fixed_qparams_qspec_observer_dedup, test/test_quantization.py::TestQuantizePT2E::test_fixed_qparams_qspec_qat, test/test_quantization.py::TestQuantizePT2E::test_fold_quantize, test/test_quantization.py::TestQuantizePT2E::test_fold_quantize_per_channel, test/test_quantization.py::TestQuantizePT2E::test_groupwise_per_channel_quant, test/test_quantization.py::TestQuantizePT2E::test_move_exported_model_bn_device_cpu, test/test_quantization.py::TestQuantizePT2E::test_preserve_nn_module_stack, test/test_quantization.py::TestQuantizePT2E::test_quantization_dtype_bfloat16_float8_e5m2, test/test_quantization.py::TestQuantizePT2E::test_quantization_dtype_float32_int16, test/test_quantization.py::TestQuantizePT2E::test_reentrant, test/test_quantization.py::TestQuantizePT2E::test_save_load, test/test_quantization.py::TestQuantizePT2E::test_shared_qspec, test/test_quantization.py::TestQuantizePT2E::test_transform_for_annotation, test/test_quantization.py::TestQuantizePT2E::test_wo_annotate_conv_output_quantizer, test/test_quantization.py::TestQuantizePT2EAffineQuantization::test_dynamic_affine_act_per_channel_weights, test/test_quantization.py::TestPT2ERepresentation::test_conv2d, test/test_quantization.py::TestXNNPACKQuantizer::test_add_mul_long, test/test_quantization.py::TestXNNPACKQuantizer::test_cat_same_node, test/test_quantization.py::TestXNNPACKQuantizer::test_conv1d, test/test_quantization.py::TestXNNPACKQuantizer::test_dynamic_linear_int4_weight, test/test_quantization.py::TestXNNPACKQuantizer::test_linear_gru, test/test_quantization.py::TestXNNPACKQuantizer::test_linear_relu, test/test_quantization.py::TestXNNPACKQuantizer::test_linear_with_dynamic_shape, test/test_quantization.py::TestXNNPACKQuantizer::test_mul_and_inplace_mul, test/test_quantization.py::TestXNNPACKQuantizer::test_mul_float32_max, test/test_quantization.py::TestXNNPACKQuantizer::test_obs_sharing_ops, test/test_quantization.py::TestXNNPACKQuantizer::test_set_module_type_case_2, test/test_quantization.py::TestXNNPACKQuantizerModels::test_resnet18, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_cat_recipe_same_inputs, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_conv2d, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_filter_conv2d_recipe, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_filter_maxpool2d_recipe, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_flatten_recipe, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_linear_binary2, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_linear_binary_dynamic, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_linear_binary_qat, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_linear_binary_unary, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_linear_unary_dynamic, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_linear_unary_dynamic_qat, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_linear_unary_qat, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_lowering_to_x86, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_maxpool2d_recipe, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_qat_conv2d, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_qat_conv2d_binary, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_qat_conv2d_binary2, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_qat_conv2d_unary, test/test_quantization.py::TestQuantizePT2EX86Inductor::test_set_module_name_qconfig_with_underscores, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_conv_bn_bias_derived_qspec, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_conv_bn_fusion_literal_args, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_conv_bn_fusion_no_conv_bias, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn1d::test_qat_preserve_source_fn_stack, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn2d::test_fold_bn_erases_add_node, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn2d::test_fold_bn_erases_bn_node, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn2d::test_qat_conv_bn_fusion_literal_args, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn2d::test_qat_conv_bn_fusion_no_conv_bias, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn2d::test_qat_conv_bn_per_channel_weight_bias, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn2d::test_qat_conv_transpose_bn, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn2d::test_qat_per_channel_weight_custom_dtype, test/test_quantization.py::TestQuantizePT2EQAT_ConvBn2d::test_qat_preserve_source_fn_stack, test/test_quantization.py::TestFXGraphMatcher::test_dict_return_type, test/test_quantization.py::TestFXGraphMatcher::test_op_relationship_mapping, test/test_quantization.py::TestFXGraphMatcher::test_simple_fun, test/test_quantization.py::TestFXGraphMatcher::test_simple_fusion, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_add_shadow_loggers_fun_qat, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_extract_weights_conv_fun_ptq, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_int8_shadows_int8_mod, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_fp16_activations, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_fp16_shadow_activations, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_fp16_vs_linear_fp16_shadow_activations, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_fp16_weights, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_linear_kwargs_shadow, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_fqn, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_match_activations_fun_qat, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_shadow_activations_fqn, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_unsupported_op_copy_skips_shadowing, test/test_quantization.py::TestFXNumericSuiteCoreAPIs::test_user_defined_function, test/test_quantization.py::TestFXNumericSuiteNShadows::test_add_loggers_functions, test/test_quantization.py::TestFXNumericSuiteNShadows::test_extract_weights_linear, test/test_quantization.py::TestFXNumericSuiteNShadows::test_linear_mod, test/test_quantization.py::TestFXNumericSuiteNShadows::test_linear_relu_mod, test/test_quantization.py::TestFXNumericSuiteNShadows::test_logger_enabled_and_save_activations_flags, test/test_quantization.py::TestFXNumericSuiteNShadows::test_mobilenet_v2, test/test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_from_list, test/test_quantization.py::TestFXNumericSuiteNShadows::test_qconfig_multi_mapping_repr, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_activations_conv, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_activations_lstm_dynamic, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_shadow_activations_conv, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_shadow_activations_linear, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_compare_shadow_activations_lstm_dynamic, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_mobilenet_v2, test/test_quantization.py::TestFXNumericSuiteCoreAPIsModels::test_resnet18, test/test_quantization.py::TestFxModelReportDetector::test_multi_linear_model_without_per_channel, test/test_quantization.py::TestFxModelReportDetectDynamicStatic::test_nested_detection_case, test/test_quantization.py::TestFxDetectInputWeightEqualization::test_input_weight_equalization_report_gen_empty, test/test_quantization.py::TestFxDetectOutliers::test_multiple_run_consistent_spike_outlier_report_gen, test/test_quantization.py::TestFxModelReportVisualizer::test_generate_tables_no_match, test/test_quantization.py::TestEqualizeFx::test_input_weight_eq_observer, test/test_quantization.py::TestEqualizeFx::test_input_weight_equalization_prepare, test/test_quantization.py::TestEqualizeFx::test_input_weight_equalization_results, test/test_quantization.py::TestEqualizeFx::test_input_weight_equalization_weights_bias, test/test_quantization.py::TestSerialization::test_conv2d_graph, test/test_quantization.py::TestSerialization::test_conv2d_graph_v3, test/test_quantization.py::TestSerialization::test_conv2d_nobias_graph_v2, test/test_quantization.py::TestSerialization::test_linear_relu_package_quantization_transforms, test/test_quantization.py::TestQuantizeJit::test_conv_bn, test/test_quantization.py::TestQuantizeJit::test_single_linear, test/test_quantization.py::TestQuantizeJit::test_single_linear_dynamic, test/test_quantization.py::TestQuantizeJitPasses::test_conv_trace, test/test_quantization.py::TestQuantizeJitPasses::test_dedup_module_uses, test/test_quantization.py::TestQuantizeJitPasses::test_foldbn_complex_cases, test/test_quantization.py::TestQuantizeJitPasses::test_inplace_option, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_child_qconfig, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_for_if, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_for_reused_weight, test/test_quantization.py::TestQuantizeJitPasses::test_insert_observers_skip_values, test/test_quantization.py::TestQuantizeJitPasses::test_insert_quant_dequant_shared_class_type, test/test_quantization.py::TestQuantizeJitPasses::test_replicate_dequant_same_value, test/test_quantization.py::TestQuantizeJitPasses::test_replicate_dequantize, test/test_quantization.py::TestQuantizeJitOps::test_clamp, test/test_quantization.py::TestQuantizeJitOps::test_group_norm, test/test_quantization.py::TestQuantizeJitOps::test_linear, test/test_quantization.py::TestQuantizeJitOps::test_qbatch_norm_relu_BNRelu, test/test_quantization.py::TestQuantizeJitOps::test_quantized_add, test/test_quantization.py::TestQuantizeJitOps::test_quantized_conv, test/test_quantization.py::TestQuantizeJitOps::test_quantized_conv_relu, test/test_quantization.py::TestQuantizeJitOps::test_quantized_mul_scalar, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_convert_dynamic_fp16, test/test_quantization.py::TestQuantizeDynamicJitPasses::test_dynamic_shared_weights, test/test_quantization.py::TestDeprecatedJitQuantized::test_rnn_cell_quantized, test/test_quantization.py::TestDeprecatedJitQuantized::test_rnn_quantized, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_qconfig, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_quant_type, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_quantize, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_quantize_jit, test/test_quantization.py::TestAOMigrationQuantization::test_function_import_utils, test/test_quantization.py::TestAOMigrationNNQuantized::test_import_nn_qat_embedding_ops, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_linear, test/test_quantization.py::TestAOMigrationNNQuantized::test_modules_utils, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_import_nn_intrinsic_qat, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_qat_linear_relu, test/test_quantization.py::TestAOMigrationNNIntrinsic::test_modules_intrinsic_quantized_bn_relu, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_convert, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_fusion_patterns, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_graph_module, test/test_quantization.py::TestAOMigrationQuantizationFx::test_function_import_fx_pattern_utils, test/test_quantization.py::TestBitsCUDA::test_types_cuda, test/test_quantization.py::TestFloat8DtypeCUDA::test_cast_round_trip_extremes_cuda_float8_e4m3fn, test/test_quantization.py::TestFloat8DtypeCUDA::test_cast_round_trip_rte_cuda_float8_e8m0fnu, test/test_quantization.py::TestFloat8DtypeCUDA::test_cast_round_trip_soak_cuda_float8_e5m2, test/test_quantization.py::TestFloat8DtypeCUDA::test_cast_round_trip_soak_cuda_float8_e8m0fnu, test/test_quantization.py::TestFloat8DtypeCUDA::test_cat_cuda_float8_e4m3fn, test/test_quantization.py::TestFloat8DtypeCUDA::test_cat_cuda_float8_e8m0fnu, test/test_quantization.py::TestFloat8DtypeCUDA::test_creation_with_zeros_cuda_float8_e4m3fn, test/test_quantization.py::TestFloat8DtypeCUDA::test_creation_with_zeros_cuda_float8_e5m2, test/test_quantization.py::TestFloat8DtypeCUDA::test_empty_cuda_float8_e8m0fnu, test/test_quantization.py::TestFloat8DtypeCUDA::test_special_numbers_cuda_float8_e4m3fn, test/test_quantization.py::TestFloat8DtypeCUDA::test_special_numbers_cuda_float8_e5m2, test/test_quantization.py::TestFloat8DtypeCUDA::test_to_string_cuda_float8_e8m0fnu 2025-07-17T10:51:17.7266501Z 2025-07-17T10:51:18.5486058Z Running test batch 'tests to run' cost 7639.59 seconds 2025-07-17T10:51:18.5493943Z Emiting td_test_failure_stats_v2 2025-07-17T10:51:18.5500865Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1752749478_f7ff36f462fb11f086cdc6217e74b1f6 2025-07-17T10:51:18.8663142Z /var/lib/jenkins/pytorch/tools/stats/upload_metrics.py:156: UserWarning: Error uploading metric td_test_failure_stats_v2 to DynamoDB: Unable to locate credentials 2025-07-17T10:51:18.8663779Z warn(f"Error uploading metric {metric_name} to DynamoDB: {e}") 2025-07-17T10:51:18.8669719Z Emiting td_test_failure_stats_v2 2025-07-17T10:51:18.8672998Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1752749478_f82fa21c62fb11f086cdc6217e74b1f6 2025-07-17T10:51:18.8701433Z inductor/test_cudagraph_trees 1/1 failed! 2025-07-17T10:51:18.8701793Z inductor/test_cudagraph_trees_expandable_segments 1/1 failed! 2025-07-17T10:51:19.5378543Z 2025-07-17T10:51:19.5378963Z real 127m26.869s 2025-07-17T10:51:19.5379176Z user 396m8.551s 2025-07-17T10:51:19.5379334Z sys 128m1.706s 2025-07-17T10:51:19.5379487Z + sccache_epilogue 2025-07-17T10:51:19.5379709Z + echo '::group::Sccache Compilation Log' 2025-07-17T10:51:19.5380333Z ##[group]Sccache Compilation Log 2025-07-17T10:51:19.5380572Z + echo '=================== sccache compilation log ===================' 2025-07-17T10:51:19.5380862Z =================== sccache compilation log =================== 2025-07-17T10:51:19.5381239Z + python /var/lib/jenkins/pytorch/.ci/pytorch/print_sccache_log.py /var/lib/jenkins/sccache_error.log 2025-07-17T10:51:19.5503771Z + echo '=========== If your build fails, please take a look at the log above for possible reasons ===========' 2025-07-17T10:51:19.5504253Z =========== If your build fails, please take a look at the log above for possible reasons =========== 2025-07-17T10:51:19.5504560Z + sccache --show-stats 2025-07-17T10:51:19.5541144Z Compile requests 8049 2025-07-17T10:51:19.5541458Z Compile requests executed 346 2025-07-17T10:51:19.5541679Z Cache hits 38 2025-07-17T10:51:19.5541906Z Cache hits (C/C++) 38 2025-07-17T10:51:19.5542098Z Cache misses 308 2025-07-17T10:51:19.5542287Z Cache misses (C/C++) 302 2025-07-17T10:51:19.5542528Z Cache misses (HIP) 6 2025-07-17T10:51:19.5542737Z Cache hits rate 10.98 % 2025-07-17T10:51:19.5542942Z Cache hits rate (C/C++) 11.18 % 2025-07-17T10:51:19.5543140Z Cache hits rate (HIP) 0.00 % 2025-07-17T10:51:19.5543337Z Cache timeouts 0 2025-07-17T10:51:19.5543534Z Cache read errors 0 2025-07-17T10:51:19.5543715Z Forced recaches 0 2025-07-17T10:51:19.5543908Z Cache write errors 0 2025-07-17T10:51:19.5544099Z Cache errors 0 2025-07-17T10:51:19.5544337Z Compilations 308 2025-07-17T10:51:19.5544528Z Compilation failures 0 2025-07-17T10:51:19.5544721Z Non-cacheable compilations 0 2025-07-17T10:51:19.5544928Z Non-cacheable calls 207 2025-07-17T10:51:19.5545158Z Non-compilation calls 7496 2025-07-17T10:51:19.5545351Z Unsupported compiler calls 0 2025-07-17T10:51:19.5545551Z Average cache write 0.000 s 2025-07-17T10:51:19.5545755Z Average compiler 2.827 s 2025-07-17T10:51:19.5545945Z Average cache read hit 0.000 s 2025-07-17T10:51:19.5546151Z Failed distributed compilations 0 2025-07-17T10:51:19.5546294Z 2025-07-17T10:51:19.5546395Z Non-cacheable reasons: 2025-07-17T10:51:19.5546600Z unknown source language 154 2025-07-17T10:51:19.5546788Z -E 53 2025-07-17T10:51:19.5546921Z 2025-07-17T10:51:19.5547484Z Cache location Local disk: "/var/lib/jenkins/.cache/sccache" 2025-07-17T10:51:19.5547872Z Use direct/preprocessor mode? yes 2025-07-17T10:51:19.5548077Z Version (client) 0.10.0 2025-07-17T10:51:19.5548278Z Cache size 19 MiB 2025-07-17T10:51:19.5548487Z Max cache size 10 GiB 2025-07-17T10:51:19.5548693Z + sccache --stop-server 2025-07-17T10:51:19.5566744Z Stopping sccache server... 2025-07-17T10:51:19.5570847Z Compile requests 8049 2025-07-17T10:51:19.5571116Z Compile requests executed 346 2025-07-17T10:51:19.5571306Z Cache hits 38 2025-07-17T10:51:19.5571488Z Cache hits (C/C++) 38 2025-07-17T10:51:19.5571661Z Cache misses 308 2025-07-17T10:51:19.5571845Z Cache misses (C/C++) 302 2025-07-17T10:51:19.5572029Z Cache misses (HIP) 6 2025-07-17T10:51:19.5572221Z Cache hits rate 10.98 % 2025-07-17T10:51:19.5572423Z Cache hits rate (C/C++) 11.18 % 2025-07-17T10:51:19.5572731Z Cache hits rate (HIP) 0.00 % 2025-07-17T10:51:19.5572920Z Cache timeouts 0 2025-07-17T10:51:19.5573104Z Cache read errors 0 2025-07-17T10:51:19.5573285Z Forced recaches 0 2025-07-17T10:51:19.5573470Z Cache write errors 0 2025-07-17T10:51:19.5573647Z Cache errors 0 2025-07-17T10:51:19.5573827Z Compilations 308 2025-07-17T10:51:19.5574048Z Compilation failures 0 2025-07-17T10:51:19.5574270Z Non-cacheable compilations 0 2025-07-17T10:51:19.5574462Z Non-cacheable calls 207 2025-07-17T10:51:19.5574652Z Non-compilation calls 7496 2025-07-17T10:51:19.5574886Z Unsupported compiler calls 0 2025-07-17T10:51:19.5575084Z Average cache write 0.000 s 2025-07-17T10:51:19.5575279Z Average compiler 2.827 s 2025-07-17T10:51:19.5575481Z Average cache read hit 0.000 s 2025-07-17T10:51:19.5575686Z Failed distributed compilations 0 2025-07-17T10:51:19.5575811Z 2025-07-17T10:51:19.5575883Z Non-cacheable reasons: 2025-07-17T10:51:19.5576056Z unknown source language 154 2025-07-17T10:51:19.5576236Z -E 53 2025-07-17T10:51:19.5576359Z 2025-07-17T10:51:19.5576481Z Cache location Local disk: "/var/lib/jenkins/.cache/sccache" 2025-07-17T10:51:19.5576749Z Use direct/preprocessor mode? yes 2025-07-17T10:51:19.5576960Z Version (client) 0.10.0 2025-07-17T10:51:19.5577161Z Cache size 19 MiB 2025-07-17T10:51:19.5577360Z Max cache size 10 GiB 2025-07-17T10:51:19.5577572Z + echo ::endgroup:: 2025-07-17T10:51:19.5577882Z ##[endgroup] 2025-07-17T10:51:19.5628283Z ##[error]Process completed with exit code 1. 2025-07-17T10:51:19.5660453Z ##[group]Run # copy test results back to the mounted workspace, needed sudo, resulting permissions were correct 2025-07-17T10:51:19.5661142Z # copy test results back to the mounted workspace, needed sudo, resulting permissions were correct 2025-07-17T10:51:19.5661868Z docker exec -t "dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571" sh -c "cd ../pytorch && sudo cp -R test/test-reports ../workspace/test" 2025-07-17T10:51:19.5674497Z shell: /usr/bin/bash -e {0} 2025-07-17T10:51:19.5674790Z env: 2025-07-17T10:51:19.5675053Z GIT_DEFAULT_BRANCH: main 2025-07-17T10:51:19.5675404Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T10:51:19.5675797Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T10:51:19.5676202Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T10:51:19.5677123Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T10:51:19.5678091Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T10:51:19.5678377Z AWS_REGION: us-east-1 2025-07-17T10:51:19.5678704Z AWS_ACCESS_KEY_ID: *** 2025-07-17T10:51:19.5679087Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T10:51:19.5683026Z AWS_SESSION_TOKEN: *** 2025-07-17T10:51:19.5683434Z CONTAINER_NAME: dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571 2025-07-17T10:51:19.5683856Z ##[endgroup] 2025-07-17T10:51:19.7035556Z ##[group]Run docker exec -t "dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571" sh -c "sudo chown -R 1001:1001 test" 2025-07-17T10:51:19.7036384Z docker exec -t "dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571" sh -c "sudo chown -R 1001:1001 test" 2025-07-17T10:51:19.7048880Z shell: /usr/bin/bash -e {0} 2025-07-17T10:51:19.7049235Z env: 2025-07-17T10:51:19.7049491Z GIT_DEFAULT_BRANCH: main 2025-07-17T10:51:19.7049826Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T10:51:19.7050417Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T10:51:19.7050905Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T10:51:19.7051742Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T10:51:19.7052608Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T10:51:19.7052935Z AWS_REGION: us-east-1 2025-07-17T10:51:19.7053272Z AWS_ACCESS_KEY_ID: *** 2025-07-17T10:51:19.7053618Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T10:51:19.7057538Z AWS_SESSION_TOKEN: *** 2025-07-17T10:51:19.7057921Z CONTAINER_NAME: dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571 2025-07-17T10:51:19.7058341Z ##[endgroup] 2025-07-17T10:51:19.9382148Z ##[group]Run cat test/**/*_toprint.log || true 2025-07-17T10:51:19.9382437Z cat test/**/*_toprint.log || true 2025-07-17T10:51:19.9395115Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-07-17T10:51:19.9395465Z env: 2025-07-17T10:51:19.9395642Z GIT_DEFAULT_BRANCH: main 2025-07-17T10:51:19.9395864Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T10:51:19.9396157Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T10:51:19.9396426Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T10:51:19.9397105Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T10:51:19.9397781Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T10:51:19.9397975Z AWS_REGION: us-east-1 2025-07-17T10:51:19.9398239Z AWS_ACCESS_KEY_ID: *** 2025-07-17T10:51:19.9398489Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T10:51:19.9402273Z AWS_SESSION_TOKEN: *** 2025-07-17T10:51:19.9402569Z CONTAINER_NAME: dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571 2025-07-17T10:51:19.9402871Z ##[endgroup] 2025-07-17T10:51:19.9534552Z cat: 'test/**/*_toprint.log': No such file or directory 2025-07-17T10:51:19.9645948Z Prepare all required actions 2025-07-17T10:51:19.9646389Z Getting action download info 2025-07-17T10:51:20.1880990Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-07-17T10:51:22.6162418Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-07-17T10:51:23.1774732Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-07-17T10:51:23.1775055Z with: 2025-07-17T10:51:23.1775259Z use-gha: true 2025-07-17T10:51:23.1775534Z file-suffix: test-default-2-6-linux.rocm.gpu.mi300.2_46160759546 2025-07-17T10:51:23.1775809Z s3-bucket: gha-artifacts 2025-07-17T10:51:23.1775984Z env: 2025-07-17T10:51:23.1776123Z GIT_DEFAULT_BRANCH: main 2025-07-17T10:51:23.1776345Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T10:51:23.1776803Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T10:51:23.1777102Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T10:51:23.1777810Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T10:51:23.1778477Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T10:51:23.1778666Z AWS_REGION: us-east-1 2025-07-17T10:51:23.1778889Z AWS_ACCESS_KEY_ID: *** 2025-07-17T10:51:23.1779142Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T10:51:23.1783022Z AWS_SESSION_TOKEN: *** 2025-07-17T10:51:23.1783323Z CONTAINER_NAME: dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571 2025-07-17T10:51:23.1783635Z ##[endgroup] 2025-07-17T10:51:23.1852920Z ##[group]Run actions/upload-artifact@v4 2025-07-17T10:51:23.1853364Z with: 2025-07-17T10:51:23.1853713Z name: test-jsons-runattempt1-test-default-2-6-linux.rocm.gpu.mi300.2_46160759546.zip 2025-07-17T10:51:23.1854044Z retention-days: 14 2025-07-17T10:51:23.1854218Z if-no-files-found: warn 2025-07-17T10:51:23.1854414Z path: test/**/*.json 2025-07-17T10:51:23.1854584Z compression-level: 6 2025-07-17T10:51:23.1854745Z overwrite: false 2025-07-17T10:51:23.1854910Z include-hidden-files: false 2025-07-17T10:51:23.1855091Z env: 2025-07-17T10:51:23.1855246Z GIT_DEFAULT_BRANCH: main 2025-07-17T10:51:23.1855475Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T10:51:23.1855779Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T10:51:23.1856062Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T10:51:23.1856770Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T10:51:23.1857459Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T10:51:23.1857657Z AWS_REGION: us-east-1 2025-07-17T10:51:23.1857880Z AWS_ACCESS_KEY_ID: *** 2025-07-17T10:51:23.1858144Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T10:51:23.1862035Z AWS_SESSION_TOKEN: *** 2025-07-17T10:51:23.1862332Z CONTAINER_NAME: dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571 2025-07-17T10:51:23.1862644Z ##[endgroup] 2025-07-17T10:51:23.7844023Z With the provided path, there will be 6 files uploaded 2025-07-17T10:51:23.7848125Z Artifact name is valid! 2025-07-17T10:51:23.7849721Z Root directory input is valid! 2025-07-17T10:51:25.8418767Z Beginning upload of artifact content to blob storage 2025-07-17T10:51:26.0626286Z Uploaded bytes 45697 2025-07-17T10:51:26.1061905Z Finished uploading artifact content to blob storage! 2025-07-17T10:51:26.1063423Z SHA256 digest of uploaded artifact zip is 58c5ad4ee7fa1fb6e44c86da8a59e0334cfb66747693aae6ef207e4c8caca51d 2025-07-17T10:51:26.1064651Z Finalizing artifact upload 2025-07-17T10:51:26.2068001Z Artifact test-jsons-runattempt1-test-default-2-6-linux.rocm.gpu.mi300.2_46160759546.zip.zip successfully finalized. Artifact ID 3553637715 2025-07-17T10:51:26.2068837Z Artifact test-jsons-runattempt1-test-default-2-6-linux.rocm.gpu.mi300.2_46160759546.zip has been successfully uploaded! Final size is 45697 bytes. Artifact ID is 3553637715 2025-07-17T10:51:26.2074834Z Artifact download URL: https://github.com/pytorch/pytorch/actions/runs/16337959895/artifacts/3553637715 2025-07-17T10:51:26.2313817Z ##[group]Run actions/upload-artifact@v4 2025-07-17T10:51:26.2314123Z with: 2025-07-17T10:51:26.2314436Z name: test-reports-runattempt1-test-default-2-6-linux.rocm.gpu.mi300.2_46160759546.zip 2025-07-17T10:51:26.2314806Z retention-days: 14 2025-07-17T10:51:26.2314993Z if-no-files-found: ignore 2025-07-17T10:51:26.2315194Z path: test/**/*.xml test/**/*.csv 2025-07-17T10:51:26.2315415Z compression-level: 6 2025-07-17T10:51:26.2315750Z overwrite: false 2025-07-17T10:51:26.2315934Z include-hidden-files: false 2025-07-17T10:51:26.2316130Z env: 2025-07-17T10:51:26.2316298Z GIT_DEFAULT_BRANCH: main 2025-07-17T10:51:26.2316531Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T10:51:26.2316832Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T10:51:26.2317116Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T10:51:26.2317835Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T10:51:26.2318513Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T10:51:26.2318715Z AWS_REGION: us-east-1 2025-07-17T10:51:26.2318967Z AWS_ACCESS_KEY_ID: *** 2025-07-17T10:51:26.2319237Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T10:51:26.2323165Z AWS_SESSION_TOKEN: *** 2025-07-17T10:51:26.2323467Z CONTAINER_NAME: dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571 2025-07-17T10:51:26.2323779Z ##[endgroup] 2025-07-17T10:51:26.8635331Z With the provided path, there will be 174 files uploaded 2025-07-17T10:51:26.8639434Z Artifact name is valid! 2025-07-17T10:51:26.8640210Z Root directory input is valid! 2025-07-17T10:51:31.1032712Z Beginning upload of artifact content to blob storage 2025-07-17T10:51:31.6459143Z Uploaded bytes 1269984 2025-07-17T10:51:31.6848273Z Finished uploading artifact content to blob storage! 2025-07-17T10:51:31.6849223Z SHA256 digest of uploaded artifact zip is 39f496a16a15b836c35a599d429d041813d65649abf3dfde863c00c844151eb0 2025-07-17T10:51:31.6850492Z Finalizing artifact upload 2025-07-17T10:51:31.7908769Z Artifact test-reports-runattempt1-test-default-2-6-linux.rocm.gpu.mi300.2_46160759546.zip.zip successfully finalized. Artifact ID 3553638378 2025-07-17T10:51:31.7909611Z Artifact test-reports-runattempt1-test-default-2-6-linux.rocm.gpu.mi300.2_46160759546.zip has been successfully uploaded! Final size is 1269984 bytes. Artifact ID is 3553638378 2025-07-17T10:51:31.7915648Z Artifact download URL: https://github.com/pytorch/pytorch/actions/runs/16337959895/artifacts/3553638378 2025-07-17T10:51:31.8107180Z ##[group]Run actions/upload-artifact@v4 2025-07-17T10:51:31.8107421Z with: 2025-07-17T10:51:31.8107703Z name: logs-runattempt1-test-default-2-6-linux.rocm.gpu.mi300.2_46160759546.zip 2025-07-17T10:51:31.8108020Z retention-days: 14 2025-07-17T10:51:31.8108206Z if-no-files-found: ignore 2025-07-17T10:51:31.8108428Z path: usage_log.txt test/**/*.log 2025-07-17T10:51:31.8108645Z compression-level: 6 2025-07-17T10:51:31.8108830Z overwrite: false 2025-07-17T10:51:31.8109011Z include-hidden-files: false 2025-07-17T10:51:31.8109200Z env: 2025-07-17T10:51:31.8109354Z GIT_DEFAULT_BRANCH: main 2025-07-17T10:51:31.8109595Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T10:51:31.8109911Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T10:51:31.8110201Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T10:51:31.8111223Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T10:51:31.8111927Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T10:51:31.8112126Z AWS_REGION: us-east-1 2025-07-17T10:51:31.8112364Z AWS_ACCESS_KEY_ID: *** 2025-07-17T10:51:31.8112633Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T10:51:31.8116514Z AWS_SESSION_TOKEN: *** 2025-07-17T10:51:31.8116811Z CONTAINER_NAME: dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571 2025-07-17T10:51:31.8117114Z ##[endgroup] 2025-07-17T10:51:32.4458088Z Multiple search paths detected. Calculating the least common ancestor of all paths 2025-07-17T10:51:32.4459352Z The least common ancestor is /home/runner/_work/pytorch/pytorch. This will be the root directory of the artifact 2025-07-17T10:51:32.4460257Z With the provided path, there will be 170 files uploaded 2025-07-17T10:51:32.4464849Z Artifact name is valid! 2025-07-17T10:51:32.4465580Z Root directory input is valid! 2025-07-17T10:51:37.7511789Z Beginning upload of artifact content to blob storage 2025-07-17T10:51:38.2241869Z Uploaded bytes 746639 2025-07-17T10:51:38.2622612Z Finished uploading artifact content to blob storage! 2025-07-17T10:51:38.2624090Z SHA256 digest of uploaded artifact zip is 0a4cec2ff20b7145dcd38e45df19969ee9fa874e51d311744fa1510d86f4c300 2025-07-17T10:51:38.2625324Z Finalizing artifact upload 2025-07-17T10:51:38.4312993Z Artifact logs-runattempt1-test-default-2-6-linux.rocm.gpu.mi300.2_46160759546.zip.zip successfully finalized. Artifact ID 3553639160 2025-07-17T10:51:38.4313778Z Artifact logs-runattempt1-test-default-2-6-linux.rocm.gpu.mi300.2_46160759546.zip has been successfully uploaded! Final size is 746639 bytes. Artifact ID is 3553639160 2025-07-17T10:51:38.4320189Z Artifact download URL: https://github.com/pytorch/pytorch/actions/runs/16337959895/artifacts/3553639160 2025-07-17T10:51:38.4506754Z ##[group]Run # shellcheck disable=SC2156 2025-07-17T10:51:38.4507062Z # shellcheck disable=SC2156 2025-07-17T10:51:38.4507456Z find . -iname "core.[1-9]*" -exec docker exec "${CONTAINER_NAME}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-07-17T10:51:38.4520107Z shell: /usr/bin/bash -e {0} 2025-07-17T10:51:38.4520332Z env: 2025-07-17T10:51:38.4520507Z GIT_DEFAULT_BRANCH: main 2025-07-17T10:51:38.4520788Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T10:51:38.4521115Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T10:51:38.4521422Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T10:51:38.4522140Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T10:51:38.4522875Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T10:51:38.4523101Z AWS_REGION: us-east-1 2025-07-17T10:51:38.4523370Z AWS_ACCESS_KEY_ID: *** 2025-07-17T10:51:38.4523654Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T10:51:38.4527487Z AWS_SESSION_TOKEN: *** 2025-07-17T10:51:38.4527810Z CONTAINER_NAME: dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571 2025-07-17T10:51:38.4528129Z ##[endgroup] 2025-07-17T10:51:39.3964459Z ##[group]Run actions/upload-artifact@ea165f8d65b6e75b540449e92b4886f43607fa02 2025-07-17T10:51:39.3964799Z with: 2025-07-17T10:51:39.3965021Z name: coredumps-default-2-6-linux.rocm.gpu.mi300.2 2025-07-17T10:51:39.3965283Z retention-days: 14 2025-07-17T10:51:39.3965485Z if-no-files-found: ignore 2025-07-17T10:51:39.3965689Z path: ./**/core.[1-9]* 2025-07-17T10:51:39.3965886Z compression-level: 6 2025-07-17T10:51:39.3966092Z overwrite: false 2025-07-17T10:51:39.3966285Z include-hidden-files: false 2025-07-17T10:51:39.3966482Z env: 2025-07-17T10:51:39.3966655Z GIT_DEFAULT_BRANCH: main 2025-07-17T10:51:39.3966901Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-07-17T10:51:39.3967217Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-07-17T10:51:39.3967515Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-07-17T10:51:39.3968276Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD168 --device /dev/dri/renderD176 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-07-17T10:51:39.3968959Z AWS_DEFAULT_REGION: us-east-1 2025-07-17T10:51:39.3969177Z AWS_REGION: us-east-1 2025-07-17T10:51:39.3969437Z AWS_ACCESS_KEY_ID: *** 2025-07-17T10:51:39.3969724Z AWS_SECRET_ACCESS_KEY: *** 2025-07-17T10:51:39.3973568Z AWS_SESSION_TOKEN: *** 2025-07-17T10:51:39.3973886Z CONTAINER_NAME: dfdad76078bb4a53f4a87514eb17e27be5c69a8cd43bf41ec170471a505b0571 2025-07-17T10:51:39.3974343Z ##[endgroup] 2025-07-17T10:51:45.4416954Z No files were found with the provided path: ./**/core.[1-9]*. No artifacts will be uploaded. 2025-07-17T10:51:45.4638130Z Post job cleanup. 2025-07-17T10:51:45.4929487Z Logging out of registry 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-07-17T10:51:45.5261958Z Post job cleanup. 2025-07-17T10:51:45.6302188Z Post job cleanup. 2025-07-17T10:51:45.6345078Z Post job cleanup. 2025-07-17T10:51:45.7132789Z [command]/usr/bin/git version 2025-07-17T10:51:45.7163098Z git version 2.50.1 2025-07-17T10:51:45.7189616Z Copying '/home/runner/.gitconfig' to '/home/runner/_work/_temp/915ed596-2bd5-4bc6-8b77-41af933d2c4d/.gitconfig' 2025-07-17T10:51:45.7199844Z Temporarily overriding HOME='/home/runner/_work/_temp/915ed596-2bd5-4bc6-8b77-41af933d2c4d' before making global git config changes 2025-07-17T10:51:45.7200434Z Adding repository directory to the temporary git global config as a safe directory 2025-07-17T10:51:45.7203090Z [command]/usr/bin/git config --global --add safe.directory /home/runner/_work/pytorch/pytorch 2025-07-17T10:51:45.7241402Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-07-17T10:51:45.7268872Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-07-17T10:51:45.7534353Z Entering 'android/libs/fbjni' 2025-07-17T10:51:45.7579129Z Entering 'third_party/FP16' 2025-07-17T10:51:45.7623577Z Entering 'third_party/FXdiv' 2025-07-17T10:51:45.7665270Z Entering 'third_party/NNPACK' 2025-07-17T10:51:45.7708913Z Entering 'third_party/NVTX' 2025-07-17T10:51:45.7753118Z Entering 'third_party/VulkanMemoryAllocator' 2025-07-17T10:51:45.7795598Z Entering 'third_party/XNNPACK' 2025-07-17T10:51:45.7854699Z Entering 'third_party/aiter' 2025-07-17T10:51:45.7897197Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-07-17T10:51:45.7952171Z Entering 'third_party/benchmark' 2025-07-17T10:51:45.7996954Z Entering 'third_party/composable_kernel' 2025-07-17T10:51:45.8045251Z Entering 'third_party/cpp-httplib' 2025-07-17T10:51:45.8090074Z Entering 'third_party/cpuinfo' 2025-07-17T10:51:45.8135278Z Entering 'third_party/cudnn_frontend' 2025-07-17T10:51:45.8180820Z Entering 'third_party/cutlass' 2025-07-17T10:51:45.8252917Z Entering 'third_party/fbgemm' 2025-07-17T10:51:45.8314761Z Entering 'third_party/fbgemm/external/asmjit' 2025-07-17T10:51:45.8364675Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-07-17T10:51:45.8427371Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-07-17T10:51:45.8479362Z Entering 'third_party/fbgemm/external/cutlass' 2025-07-17T10:51:45.8536843Z Entering 'third_party/fbgemm/external/googletest' 2025-07-17T10:51:45.8587816Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-07-17T10:51:45.8638091Z Entering 'third_party/fbgemm/external/json' 2025-07-17T10:51:45.8692814Z Entering 'third_party/flash-attention' 2025-07-17T10:51:45.8750109Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-07-17T10:51:45.8807814Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-07-17T10:51:45.8866681Z Entering 'third_party/flatbuffers' 2025-07-17T10:51:45.8922271Z Entering 'third_party/fmt' 2025-07-17T10:51:45.8973790Z Entering 'third_party/gemmlowp/gemmlowp' 2025-07-17T10:51:45.9026941Z Entering 'third_party/gloo' 2025-07-17T10:51:45.9077908Z Entering 'third_party/googletest' 2025-07-17T10:51:45.9130913Z Entering 'third_party/ideep' 2025-07-17T10:51:45.9182687Z Entering 'third_party/ideep/mkl-dnn' 2025-07-17T10:51:45.9243155Z Entering 'third_party/ittapi' 2025-07-17T10:51:45.9309841Z Entering 'third_party/kineto' 2025-07-17T10:51:45.9362933Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-07-17T10:51:45.9414778Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-07-17T10:51:45.9464271Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-07-17T10:51:45.9514069Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-07-17T10:51:45.9564529Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-07-17T10:51:45.9619331Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-07-17T10:51:45.9678663Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-07-17T10:51:45.9729000Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-07-17T10:51:45.9779660Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-07-17T10:51:45.9834250Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-07-17T10:51:45.9893041Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-07-17T10:51:45.9942692Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-07-17T10:51:45.9999720Z Entering 'third_party/kleidiai' 2025-07-17T10:51:46.0051214Z Entering 'third_party/mimalloc' 2025-07-17T10:51:46.0101895Z Entering 'third_party/nlohmann' 2025-07-17T10:51:46.0159891Z Entering 'third_party/onnx' 2025-07-17T10:51:46.0229737Z Entering 'third_party/onnx/third_party/pybind11' 2025-07-17T10:51:46.0285361Z Entering 'third_party/opentelemetry-cpp' 2025-07-17T10:51:46.0338268Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-07-17T10:51:46.0390853Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-07-17T10:51:46.0446982Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-07-17T10:51:46.0501516Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-07-17T10:51:46.0554149Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-07-17T10:51:46.0605571Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-07-17T10:51:46.0657334Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-07-17T10:51:46.0710696Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-07-17T10:51:46.0764909Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-07-17T10:51:46.0817665Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-07-17T10:51:46.0889344Z Entering 'third_party/pocketfft' 2025-07-17T10:51:46.0942453Z Entering 'third_party/protobuf' 2025-07-17T10:51:46.0999422Z Entering 'third_party/protobuf/third_party/benchmark' 2025-07-17T10:51:46.1049246Z Entering 'third_party/protobuf/third_party/googletest' 2025-07-17T10:51:46.1102967Z Entering 'third_party/psimd' 2025-07-17T10:51:46.1157447Z Entering 'third_party/pthreadpool' 2025-07-17T10:51:46.1210247Z Entering 'third_party/pybind11' 2025-07-17T10:51:46.1262513Z Entering 'third_party/python-peachpy' 2025-07-17T10:51:46.1315297Z Entering 'third_party/sleef' 2025-07-17T10:51:46.1366520Z Entering 'third_party/tensorpipe' 2025-07-17T10:51:46.1417061Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-07-17T10:51:46.1470197Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-07-17T10:51:46.1521989Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-07-17T10:51:46.1572550Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-07-17T10:51:46.1628302Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-07-17T10:51:46.1703448Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-07-17T10:51:46.1731000Z http.https://github.com/.extraheader 2025-07-17T10:51:46.1741629Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-07-17T10:51:46.1776752Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-07-17T10:51:46.2069171Z Entering 'android/libs/fbjni' 2025-07-17T10:51:46.2099746Z http.https://github.com/.extraheader 2025-07-17T10:51:46.2143553Z Entering 'third_party/FP16' 2025-07-17T10:51:46.2172946Z http.https://github.com/.extraheader 2025-07-17T10:51:46.2214420Z Entering 'third_party/FXdiv' 2025-07-17T10:51:46.2244632Z http.https://github.com/.extraheader 2025-07-17T10:51:46.2284312Z Entering 'third_party/NNPACK' 2025-07-17T10:51:46.2314314Z http.https://github.com/.extraheader 2025-07-17T10:51:46.2357435Z Entering 'third_party/NVTX' 2025-07-17T10:51:46.2386071Z http.https://github.com/.extraheader 2025-07-17T10:51:46.2430847Z Entering 'third_party/VulkanMemoryAllocator' 2025-07-17T10:51:46.2460255Z http.https://github.com/.extraheader 2025-07-17T10:51:46.2500682Z Entering 'third_party/XNNPACK' 2025-07-17T10:51:46.2532221Z http.https://github.com/.extraheader 2025-07-17T10:51:46.2587800Z Entering 'third_party/aiter' 2025-07-17T10:51:46.2616321Z http.https://github.com/.extraheader 2025-07-17T10:51:46.2660045Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-07-17T10:51:46.2688313Z http.https://github.com/.extraheader 2025-07-17T10:51:46.2742255Z Entering 'third_party/benchmark' 2025-07-17T10:51:46.2772757Z http.https://github.com/.extraheader 2025-07-17T10:51:46.2815629Z Entering 'third_party/composable_kernel' 2025-07-17T10:51:46.2845402Z http.https://github.com/.extraheader 2025-07-17T10:51:46.2896667Z Entering 'third_party/cpp-httplib' 2025-07-17T10:51:46.2926393Z http.https://github.com/.extraheader 2025-07-17T10:51:46.2968204Z Entering 'third_party/cpuinfo' 2025-07-17T10:51:46.2997206Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3039647Z Entering 'third_party/cudnn_frontend' 2025-07-17T10:51:46.3067015Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3109121Z Entering 'third_party/cutlass' 2025-07-17T10:51:46.3137022Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3189509Z Entering 'third_party/fbgemm' 2025-07-17T10:51:46.3219993Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3264748Z Entering 'third_party/fbgemm/external/asmjit' 2025-07-17T10:51:46.3295137Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3334557Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-07-17T10:51:46.3365525Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3413554Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-07-17T10:51:46.3443280Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3494707Z Entering 'third_party/fbgemm/external/cutlass' 2025-07-17T10:51:46.3515200Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3565948Z Entering 'third_party/fbgemm/external/googletest' 2025-07-17T10:51:46.3597203Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3637927Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-07-17T10:51:46.3668323Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3711195Z Entering 'third_party/fbgemm/external/json' 2025-07-17T10:51:46.3740791Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3786334Z Entering 'third_party/flash-attention' 2025-07-17T10:51:46.3818080Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3861976Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-07-17T10:51:46.3892111Z http.https://github.com/.extraheader 2025-07-17T10:51:46.3941833Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-07-17T10:51:46.3971032Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4025792Z Entering 'third_party/flatbuffers' 2025-07-17T10:51:46.4054466Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4100377Z Entering 'third_party/fmt' 2025-07-17T10:51:46.4131301Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4173250Z Entering 'third_party/gemmlowp/gemmlowp' 2025-07-17T10:51:46.4202796Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4243679Z Entering 'third_party/gloo' 2025-07-17T10:51:46.4272885Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4315070Z Entering 'third_party/googletest' 2025-07-17T10:51:46.4343841Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4384827Z Entering 'third_party/ideep' 2025-07-17T10:51:46.4415163Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4455804Z Entering 'third_party/ideep/mkl-dnn' 2025-07-17T10:51:46.4484828Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4535767Z Entering 'third_party/ittapi' 2025-07-17T10:51:46.4565575Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4608406Z Entering 'third_party/kineto' 2025-07-17T10:51:46.4637974Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4680434Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-07-17T10:51:46.4709484Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4753401Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-07-17T10:51:46.4784817Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4830381Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-07-17T10:51:46.4859176Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4903787Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-07-17T10:51:46.4933374Z http.https://github.com/.extraheader 2025-07-17T10:51:46.4976945Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-07-17T10:51:46.5005866Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5050179Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-07-17T10:51:46.5079911Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5127590Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-07-17T10:51:46.5154555Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5199573Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-07-17T10:51:46.5227985Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5273623Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-07-17T10:51:46.5303918Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5349504Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-07-17T10:51:46.5379117Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5428257Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-07-17T10:51:46.5459140Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5501348Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-07-17T10:51:46.5530913Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5577195Z Entering 'third_party/kleidiai' 2025-07-17T10:51:46.5606175Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5649913Z Entering 'third_party/mimalloc' 2025-07-17T10:51:46.5680008Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5721533Z Entering 'third_party/nlohmann' 2025-07-17T10:51:46.5751255Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5798831Z Entering 'third_party/onnx' 2025-07-17T10:51:46.5828299Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5885686Z Entering 'third_party/onnx/third_party/pybind11' 2025-07-17T10:51:46.5913570Z http.https://github.com/.extraheader 2025-07-17T10:51:46.5961912Z Entering 'third_party/opentelemetry-cpp' 2025-07-17T10:51:46.5991094Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6032781Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-07-17T10:51:46.6061870Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6105029Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-07-17T10:51:46.6133212Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6177339Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-07-17T10:51:46.6205701Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6249477Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-07-17T10:51:46.6281592Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6326586Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-07-17T10:51:46.6355630Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6398810Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-07-17T10:51:46.6427451Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6469583Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-07-17T10:51:46.6499594Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6541852Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-07-17T10:51:46.6571912Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6615321Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-07-17T10:51:46.6644542Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6689882Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-07-17T10:51:46.6718289Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6779527Z Entering 'third_party/pocketfft' 2025-07-17T10:51:46.6809876Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6854251Z Entering 'third_party/protobuf' 2025-07-17T10:51:46.6887242Z http.https://github.com/.extraheader 2025-07-17T10:51:46.6931402Z Entering 'third_party/protobuf/third_party/benchmark' 2025-07-17T10:51:46.6960184Z http.https://github.com/.extraheader 2025-07-17T10:51:46.7003537Z Entering 'third_party/protobuf/third_party/googletest' 2025-07-17T10:51:46.7033079Z http.https://github.com/.extraheader 2025-07-17T10:51:46.7081239Z Entering 'third_party/psimd' 2025-07-17T10:51:46.7112172Z http.https://github.com/.extraheader 2025-07-17T10:51:46.7154133Z Entering 'third_party/pthreadpool' 2025-07-17T10:51:46.7182975Z http.https://github.com/.extraheader 2025-07-17T10:51:46.7224625Z Entering 'third_party/pybind11' 2025-07-17T10:51:46.7256252Z http.https://github.com/.extraheader 2025-07-17T10:51:46.7295589Z Entering 'third_party/python-peachpy' 2025-07-17T10:51:46.7325075Z http.https://github.com/.extraheader 2025-07-17T10:51:46.7367239Z Entering 'third_party/sleef' 2025-07-17T10:51:46.7396585Z http.https://github.com/.extraheader 2025-07-17T10:51:46.7439484Z Entering 'third_party/tensorpipe' 2025-07-17T10:51:46.7469781Z http.https://github.com/.extraheader 2025-07-17T10:51:46.7512343Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-07-17T10:51:46.7541801Z http.https://github.com/.extraheader 2025-07-17T10:51:46.7582417Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-07-17T10:51:46.7611742Z http.https://github.com/.extraheader 2025-07-17T10:51:46.7652630Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-07-17T10:51:46.7681928Z http.https://github.com/.extraheader 2025-07-17T10:51:46.7726125Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-07-17T10:51:46.7754255Z http.https://github.com/.extraheader 2025-07-17T10:51:46.7794146Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-07-17T10:51:46.7823299Z http.https://github.com/.extraheader 2025-07-17T10:51:46.8031805Z Cleaning up orphan processes